id
stringlengths 36
36
| filename
stringclasses 310
values | label
stringclasses 13
values | text
stringlengths 0
25.7k
⌀ | caption_text
stringlengths 2
1.49k
⌀ | image
imagewidth (px) 2
1.11k
⌀ | width
int64 0
1.11k
| height
int64 0
1.42k
| dpi
int64 72
72
| mimetype
stringclasses 1
value | page_no
int64 2
305
| mime_type
stringclasses 1
value | version
stringclasses 1
value | tags
sequencelengths 0
0
| properties
null | error
stringclasses 1
value | raw_response
stringlengths 255
1.07M
| synced_at
null |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
f7aaa3c0-262d-4894-a0d2-1a4d50f0117c | 2302.06555v2.pdf | page_header | arXiv:2302.06555v2 [cs.CL] 6 Jul 2024 | null | 39 | 660 | 72 | image/png | 2 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/0", "parent": {"cref": "#/body"}, "children": [], "label": "page_header", "prov": [{"page_no": 1, "bbox": {"l": 17.23870086669922, "t": 566.97998046875, "r": 36.33979415893555, "b": 236.99996948242188, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 37]}], "orig": "arXiv:2302.06555v2 [cs.CL] 6 Jul 2024", "text": "arXiv:2302.06555v2 [cs.CL] 6 Jul 2024"} | null |
|
1efbff73-8fa6-4fe1-9456-df4f893eb0cf | 2302.06555v2.pdf | section_header | Do Vision and Language Models Share Concepts? A Vector Space Alignment Study | null | 613 | 60 | 72 | image/png | 2 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/1", "parent": {"cref": "#/body"}, "children": [], "label": "section_header", "prov": [{"page_no": 1, "bbox": {"l": 145.64881896972656, "t": 772.0592651367188, "r": 451.7751159667969, "b": 741.8055419921875, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 76]}], "orig": "Do Vision and Language Models Share Concepts? A Vector Space Alignment Study", "text": "Do Vision and Language Models Share Concepts? A Vector Space Alignment Study", "level": 1} | null |
|
2f6fc597-e04a-4529-8017-f9e746c2bf7c | 2302.06555v2.pdf | section_header | Jiaang Li † Yova Kementchedjhieva ‡ Constanza Fierro † Anders Søgaard † | null | 757 | 26 | 72 | image/png | 2 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/2", "parent": {"cref": "#/body"}, "children": [], "label": "section_header", "prov": [{"page_no": 1, "bbox": {"l": 110.50749206542969, "t": 728.587646484375, "r": 488.74749755859375, "b": 715.568359375, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 71]}], "orig": "Jiaang Li \u2020 Yova Kementchedjhieva \u2021 Constanza Fierro \u2020 Anders S\u00f8gaard \u2020", "text": "Jiaang Li \u2020 Yova Kementchedjhieva \u2021 Constanza Fierro \u2020 Anders S\u00f8gaard \u2020", "level": 1} | null |
|
835d05b7-a176-4446-8c03-5cfd00098e08 | 2302.06555v2.pdf | text | † University of Copenhagen | null | 267 | 28 | 72 | image/png | 2 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/3", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 1, "bbox": {"l": 233.31304931640625, "t": 700.5826416015625, "r": 366.739990234375, "b": 686.6187744140625, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 26]}], "orig": "\u2020 University of Copenhagen", "text": "\u2020 University of Copenhagen"} | null |
|
9960238e-0bda-414a-b41a-5179b70ccfad | 2302.06555v2.pdf | text | ‡ Mohamed bin Zayed University of Artificial Intelligence | null | 558 | 27 | 72 | image/png | 2 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/4", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 1, "bbox": {"l": 160.5359344482422, "t": 686.524658203125, "r": 439.458984375, "b": 672.8939208984375, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 57]}], "orig": "\u2021 Mohamed bin Zayed University of Artificial Intelligence", "text": "\u2021 Mohamed bin Zayed University of Artificial Intelligence"} | null |
|
d223e870-906b-4539-969c-496215626791 | 2302.06555v2.pdf | text | {jili,c.fierro,soegaard}@di.ku.dk, [email protected] | null | 990 | 25 | 72 | image/png | 2 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/5", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 1, "bbox": {"l": 52.79401397705078, "t": 671.6949462890625, "r": 547.7392578125, "b": 659.25048828125, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 69]}], "orig": "{jili,c.fierro,soegaard}@di.ku.dk, [email protected]", "text": "{jili,c.fierro,soegaard}@di.ku.dk, [email protected]"} | null |
|
7e2361f3-7c96-4be2-ad2b-02bb14073d09 | 2302.06555v2.pdf | section_header | Abstract | null | 91 | 23 | 72 | image/png | 2 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/6", "parent": {"cref": "#/body"}, "children": [], "label": "section_header", "prov": [{"page_no": 1, "bbox": {"l": 158.09361267089844, "t": 635.4051513671875, "r": 203.5260009765625, "b": 623.8914184570312, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 8]}], "orig": "Abstract", "text": "Abstract", "level": 1} | null |
|
b4c1466b-92d1-411f-8b3e-c23c96a2cd1c | 2302.06555v2.pdf | text | Large-scale pretrained language models (LMs) are said to "lack the ability to connect utterances to the world" (Bender and Koller, 2020), because they do not have "mental models of the world" (Mitchell and Krakauer, 2023). If so, one would expect LM representations to be unrelated to representations induced by vision models. We present an empirical evaluation across four families of LMs (BERT, GPT-2, OPT and LLaMA-2) and three vision model architectures (ResNet, SegFormer, and MAE). Our experiments show that LMs partially converge towards representations isomorphic to those of vision models, subject to dispersion, polysemy and frequency. This has important implications for both multi-modal processing and the LM understanding debate (Mitchell and Krakauer, 2023). 1 | null | 355 | 449 | 72 | image/png | 2 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/7", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 1, "bbox": {"l": 92.48450469970703, "t": 607.5615844726562, "r": 270.1062316894531, "b": 382.78509521484375, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 774]}], "orig": "Large-scale pretrained language models (LMs) are said to \"lack the ability to connect utterances to the world\" (Bender and Koller, 2020), because they do not have \"mental models of the world\" (Mitchell and Krakauer, 2023). If so, one would expect LM representations to be unrelated to representations induced by vision models. We present an empirical evaluation across four families of LMs (BERT, GPT-2, OPT and LLaMA-2) and three vision model architectures (ResNet, SegFormer, and MAE). Our experiments show that LMs partially converge towards representations isomorphic to those of vision models, subject to dispersion, polysemy and frequency. This has important implications for both multi-modal processing and the LM understanding debate (Mitchell and Krakauer, 2023). 1", "text": "Large-scale pretrained language models (LMs) are said to \"lack the ability to connect utterances to the world\" (Bender and Koller, 2020), because they do not have \"mental models of the world\" (Mitchell and Krakauer, 2023). If so, one would expect LM representations to be unrelated to representations induced by vision models. We present an empirical evaluation across four families of LMs (BERT, GPT-2, OPT and LLaMA-2) and three vision model architectures (ResNet, SegFormer, and MAE). Our experiments show that LMs partially converge towards representations isomorphic to those of vision models, subject to dispersion, polysemy and frequency. This has important implications for both multi-modal processing and the LM understanding debate (Mitchell and Krakauer, 2023). 1"} | null |
|
2aa84d33-241d-454f-aff6-090baeb93711 | 2302.06555v2.pdf | section_header | 1 Introduction | null | 166 | 22 | 72 | image/png | 2 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/8", "parent": {"cref": "#/body"}, "children": [], "label": "section_header", "prov": [{"page_no": 1, "bbox": {"l": 71.96525573730469, "t": 357.04547119140625, "r": 154.81365966796875, "b": 345.7883605957031, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 14]}], "orig": "1 Introduction", "text": "1 Introduction", "level": 1} | null |
|
43ebcc8a-6b10-4dc7-b1bb-1d1df53ce5e0 | 2302.06555v2.pdf | text | The debate around whether LMs can be said to understand is often portrayed as a back-and-forth between two opposing sides (Mitchell and Krakauer, 2023), but in reality, there are many positions. Some researchers have argued that LMs are 'all syntax, no semantics', i.e., that they learn form, but not meaning (Searle, 1980; Bender and Koller, 2020; Marcus et al., 2023). 2 Others have argued that LMs | null | 442 | 213 | 72 | image/png | 2 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/9", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 1, "bbox": {"l": 70.90017700195312, "t": 334.8507995605469, "r": 292.1755065917969, "b": 228.49066162109375, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 400]}], "orig": "The debate around whether LMs can be said to understand is often portrayed as a back-and-forth between two opposing sides (Mitchell and Krakauer, 2023), but in reality, there are many positions. Some researchers have argued that LMs are 'all syntax, no semantics', i.e., that they learn form, but not meaning (Searle, 1980; Bender and Koller, 2020; Marcus et al., 2023). 2 Others have argued that LMs", "text": "The debate around whether LMs can be said to understand is often portrayed as a back-and-forth between two opposing sides (Mitchell and Krakauer, 2023), but in reality, there are many positions. Some researchers have argued that LMs are 'all syntax, no semantics', i.e., that they learn form, but not meaning (Searle, 1980; Bender and Koller, 2020; Marcus et al., 2023). 2 Others have argued that LMs"} | null |
|
15bed161-2aec-4ca4-ab02-b65cd1cab586 | 2302.06555v2.pdf | text | dataset: | null | 57 | 17 | 72 | image/png | 2 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/10", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 1, "bbox": {"l": 141.05947875976562, "t": 217.11761474609375, "r": 169.70974731445312, "b": 208.6852569580078, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 8]}], "orig": "dataset:", "text": "dataset:"} | null |
|
7e34928e-b3ee-434d-94db-81cb08d3c4d6 | 2302.06555v2.pdf | text | https://github.com/ | null | 206 | 19 | 72 | image/png | 2 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/11", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 1, "bbox": {"l": 188.3842010498047, "t": 217.300048828125, "r": 291.3439636230469, "b": 208.0350341796875, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 19]}], "orig": "https://github.com/", "text": "https://github.com/"} | null |
|
9bf2f166-4937-4ace-bb94-dfb344f88b46 | 2302.06555v2.pdf | text | $^{1}$Code and jiaangli/VLCA . | null | 144 | 37 | 72 | image/png | 2 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/12", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 1, "bbox": {"l": 72.0, "t": 216.7012176513672, "r": 144.17959594726562, "b": 197.72625732421875, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 30]}], "orig": "$^{1}$Code and jiaangli/VLCA .", "text": "$^{1}$Code and jiaangli/VLCA ."} | null |
|
6321b136-a71f-471e-a05c-035bebe99d9a | 2302.06555v2.pdf | text | $^{2}$The idea that computers are 'all syntax, no semantics' can be traced back to German 17th century philosopher Leibniz's Mill Argument (Lodge and Bobro, 1998). The Mill Argument states that mental states cannot be reduced to physical states, so if the capacity to understand language requires mental states, this capacity cannot be instantiated, merely imitated, by machines. In 1980, Searle introduced an even more popular argument against the possibility of LM understanding, in the form of the so-called Chinese Room thought experiment (Searle, 1980). The Chinese Room presents an interlocutor with no prior knowledge of a foreign language, who | null | 442 | 238 | 72 | image/png | 2 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/13", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 1, "bbox": {"l": 71.13861846923828, "t": 195.3333740234375, "r": 291.759765625, "b": 76.2840576171875, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 651]}], "orig": "$^{2}$The idea that computers are 'all syntax, no semantics' can be traced back to German 17th century philosopher Leibniz's Mill Argument (Lodge and Bobro, 1998). The Mill Argument states that mental states cannot be reduced to physical states, so if the capacity to understand language requires mental states, this capacity cannot be instantiated, merely imitated, by machines. In 1980, Searle introduced an even more popular argument against the possibility of LM understanding, in the form of the so-called Chinese Room thought experiment (Searle, 1980). The Chinese Room presents an interlocutor with no prior knowledge of a foreign language, who", "text": "$^{2}$The idea that computers are 'all syntax, no semantics' can be traced back to German 17th century philosopher Leibniz's Mill Argument (Lodge and Bobro, 1998). The Mill Argument states that mental states cannot be reduced to physical states, so if the capacity to understand language requires mental states, this capacity cannot be instantiated, merely imitated, by machines. In 1980, Searle introduced an even more popular argument against the possibility of LM understanding, in the form of the so-called Chinese Room thought experiment (Searle, 1980). The Chinese Room presents an interlocutor with no prior knowledge of a foreign language, who"} | null |
|
8bb27c97-a19f-443c-b6c3-10c0c3fea2fa | 2302.06555v2.pdf | text | have inferential semantics, but not referential semantics (Rapaport, 2002; Sahlgren and Carlsson, 2021; Piantadosi and Hill, 2022), 3 whereas some have posited that a form of externalist referential semantics is possible, at least for chatbots engaged in direct conversation (Cappelen and Dever, 2021; Butlin, 2021; Mollo and Millière, 2023; Mandelkern and Linzen, 2023). Most researchers agree, however, that LMs "lack the ability to connect utterances to the world" (Bender and Koller, 2020), because they do not have "mental models of the world" (Mitchell and Krakauer, 2023). | null | 443 | 320 | 72 | image/png | 2 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/14", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 1, "bbox": {"l": 306.2083740234375, "t": 634.934326171875, "r": 527.3598022460938, "b": 474.9033508300781, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 579]}], "orig": "have inferential semantics, but not referential semantics (Rapaport, 2002; Sahlgren and Carlsson, 2021; Piantadosi and Hill, 2022), 3 whereas some have posited that a form of externalist referential semantics is possible, at least for chatbots engaged in direct conversation (Cappelen and Dever, 2021; Butlin, 2021; Mollo and Milli\u00e8re, 2023; Mandelkern and Linzen, 2023). Most researchers agree, however, that LMs \"lack the ability to connect utterances to the world\" (Bender and Koller, 2020), because they do not have \"mental models of the world\" (Mitchell and Krakauer, 2023).", "text": "have inferential semantics, but not referential semantics (Rapaport, 2002; Sahlgren and Carlsson, 2021; Piantadosi and Hill, 2022), 3 whereas some have posited that a form of externalist referential semantics is possible, at least for chatbots engaged in direct conversation (Cappelen and Dever, 2021; Butlin, 2021; Mollo and Milli\u00e8re, 2023; Mandelkern and Linzen, 2023). Most researchers agree, however, that LMs \"lack the ability to connect utterances to the world\" (Bender and Koller, 2020), because they do not have \"mental models of the world\" (Mitchell and Krakauer, 2023)."} | null |
|
78879112-8454-4241-9808-cdc6670cb00a | 2302.06555v2.pdf | text | This study provides evidence to the contrary: Language models and computer vision models (VMs) are trained on independent data sources (at least for unsupervised computer vision models). The only common source of bias is the world. If LMs and VMs exhibit similarities, it must be because they both model the world. We examine the representations learned by different LMs and VMs by measuring how similar their geometries are. We consistently find that the better the LMs are, the more they induce representations similar to those induced by computer vision models. The similarity between the two spaces is such that from a very small set of parallel examples we are able to linearly project VMs representations to the language space and retrieve highly accurate captions, as shown by the examples in Figure 1. | null | 443 | 457 | 72 | image/png | 2 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/15", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 1, "bbox": {"l": 305.9925231933594, "t": 471.6637268066406, "r": 527.4515380859375, "b": 242.90240478515625, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 809]}], "orig": "This study provides evidence to the contrary: Language models and computer vision models (VMs) are trained on independent data sources (at least for unsupervised computer vision models). The only common source of bias is the world. If LMs and VMs exhibit similarities, it must be because they both model the world. We examine the representations learned by different LMs and VMs by measuring how similar their geometries are. We consistently find that the better the LMs are, the more they induce representations similar to those induced by computer vision models. The similarity between the two spaces is such that from a very small set of parallel examples we are able to linearly project VMs representations to the language space and retrieve highly accurate captions, as shown by the examples in Figure 1.", "text": "This study provides evidence to the contrary: Language models and computer vision models (VMs) are trained on independent data sources (at least for unsupervised computer vision models). The only common source of bias is the world. If LMs and VMs exhibit similarities, it must be because they both model the world. We examine the representations learned by different LMs and VMs by measuring how similar their geometries are. We consistently find that the better the LMs are, the more they induce representations similar to those induced by computer vision models. The similarity between the two spaces is such that from a very small set of parallel examples we are able to linearly project VMs representations to the language space and retrieve highly accurate captions, as shown by the examples in Figure 1."} | null |
|
13524540-4a6d-47b7-b3ad-05ded451e35c | 2302.06555v2.pdf | text | Contributions. We present a series of evaluations of the vector spaces induced by three families of VMs and four families of LMs, i.e., a total of fourteen VMs and fourteen LMs. We show that within each family, the larger the LMs, the more their vector spaces become structurally similar to | null | 442 | 158 | 72 | image/png | 2 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/16", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 1, "bbox": {"l": 306.3404541015625, "t": 230.98822021484375, "r": 527.3585815429688, "b": 151.7589111328125, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 290]}], "orig": "Contributions. We present a series of evaluations of the vector spaces induced by three families of VMs and four families of LMs, i.e., a total of fourteen VMs and fourteen LMs. We show that within each family, the larger the LMs, the more their vector spaces become structurally similar to", "text": "Contributions. We present a series of evaluations of the vector spaces induced by three families of VMs and four families of LMs, i.e., a total of fourteen VMs and fourteen LMs. We show that within each family, the larger the LMs, the more their vector spaces become structurally similar to"} | null |
|
35f66de4-2a68-4b7f-99ab-fe619540be51 | 2302.06555v2.pdf | text | receives text messages in this language and follows a rule book to reply to the messages. The interlocutor is Searle's caricature of artificial intelligence, and is obviously, Searle claims, not endowed with meaning or understanding, but merely symbol manipulation. | null | 439 | 106 | 72 | image/png | 2 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/17", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 1, "bbox": {"l": 306.3285217285156, "t": 140.958984375, "r": 525.7665405273438, "b": 88.13725280761719, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 265]}], "orig": "receives text messages in this language and follows a rule book to reply to the messages. The interlocutor is Searle's caricature of artificial intelligence, and is obviously, Searle claims, not endowed with meaning or understanding, but merely symbol manipulation.", "text": "receives text messages in this language and follows a rule book to reply to the messages. The interlocutor is Searle's caricature of artificial intelligence, and is obviously, Searle claims, not endowed with meaning or understanding, but merely symbol manipulation."} | null |
|
cbc18cec-292c-43a7-a222-f02d37ed6869 | 2302.06555v2.pdf | text | $^{3}$See Marconi (1997) for this distinction. | null | 291 | 16 | 72 | image/png | 2 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/18", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 1, "bbox": {"l": 319.9280090332031, "t": 85.00321197509766, "r": 465.3711242675781, "b": 76.98725128173828, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 46]}], "orig": "$^{3}$See Marconi (1997) for this distinction.", "text": "$^{3}$See Marconi (1997) for this distinction."} | null |
|
a4008690-b978-45e3-ab18-4930e3416f4c | 2302.06555v2.pdf | text | those of computer vision models. This enables retrieval of language representations of images (referential semantics) with minimal supervision. Retrieval precision depends on dispersion of image and language, polysemy, and frequency, but consistently improves with language model size. We discuss the implications of the finding that language and computer vision models learn representations with similar geometries. | null | 442 | 238 | 72 | image/png | 3 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/19", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 2, "bbox": {"l": 71.13309478759766, "t": 776.2091064453125, "r": 292.083251953125, "b": 657.3726196289062, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 416]}], "orig": "those of computer vision models. This enables retrieval of language representations of images (referential semantics) with minimal supervision. Retrieval precision depends on dispersion of image and language, polysemy, and frequency, but consistently improves with language model size. We discuss the implications of the finding that language and computer vision models learn representations with similar geometries.", "text": "those of computer vision models. This enables retrieval of language representations of images (referential semantics) with minimal supervision. Retrieval precision depends on dispersion of image and language, polysemy, and frequency, but consistently improves with language model size. We discuss the implications of the finding that language and computer vision models learn representations with similar geometries."} | null |
|
7c58bc30-6e15-40ce-b04a-380f6e1ab0e2 | 2302.06555v2.pdf | section_header | 2 Related Work | null | 180 | 23 | 72 | image/png | 3 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/20", "parent": {"cref": "#/body"}, "children": [], "label": "section_header", "prov": [{"page_no": 2, "bbox": {"l": 71.47272491455078, "t": 645.0098876953125, "r": 161.3809814453125, "b": 633.350341796875, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 14]}], "orig": "2 Related Work", "text": "2 Related Work", "level": 1} | null |
|
553eacd2-92df-4108-be99-f90ab08aac40 | 2302.06555v2.pdf | text | Inspiration from cognitive science. Computational modeling is a cornerstone of cognitive science in the pursuit for a better understanding of how representations in the brain come about. As such, the field has shown a growing interest in computational representations induced with self-supervised learning (Orhan et al., 2020; Halvagal and Zenke, 2022). Cognitive scientists have also noted how the objectives of supervised language and vision models bear resemblances to predictive processing (Schrimpf et al., 2018; Goldstein et al., 2021; Caucheteux et al., 2022; Li et al., 2023) (but see Antonello and Huth (2022) for a critical discussion of such work). | null | 442 | 373 | 72 | image/png | 3 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/21", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 2, "bbox": {"l": 71.06283569335938, "t": 622.7850952148438, "r": 292.0802001953125, "b": 436.3576354980469, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 659]}], "orig": "Inspiration from cognitive science. Computational modeling is a cornerstone of cognitive science in the pursuit for a better understanding of how representations in the brain come about. As such, the field has shown a growing interest in computational representations induced with self-supervised learning (Orhan et al., 2020; Halvagal and Zenke, 2022). Cognitive scientists have also noted how the objectives of supervised language and vision models bear resemblances to predictive processing (Schrimpf et al., 2018; Goldstein et al., 2021; Caucheteux et al., 2022; Li et al., 2023) (but see Antonello and Huth (2022) for a critical discussion of such work).", "text": "Inspiration from cognitive science. Computational modeling is a cornerstone of cognitive science in the pursuit for a better understanding of how representations in the brain come about. As such, the field has shown a growing interest in computational representations induced with self-supervised learning (Orhan et al., 2020; Halvagal and Zenke, 2022). Cognitive scientists have also noted how the objectives of supervised language and vision models bear resemblances to predictive processing (Schrimpf et al., 2018; Goldstein et al., 2021; Caucheteux et al., 2022; Li et al., 2023) (but see Antonello and Huth (2022) for a critical discussion of such work)."} | null |
|
e92fdcac-17c2-416a-a570-6a4387d1f6b8 | 2302.06555v2.pdf | text | Studies have looked at the alignability of neural language representations and human brain activations, with more promising results as language models grow better at modeling language (Sassenhagen and Fiebach, 2020; Schrimpf et al., 2021). In these studies, the partial alignability of brain and model representations is interpreted as evidence that brain and models might process language in the same way (Caucheteux and King, 2022). | null | 441 | 240 | 72 | image/png | 3 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/22", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 2, "bbox": {"l": 71.25102233886719, "t": 433.5199279785156, "r": 292.1755065917969, "b": 313.6751708984375, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 434]}], "orig": "Studies have looked at the alignability of neural language representations and human brain activations, with more promising results as language models grow better at modeling language (Sassenhagen and Fiebach, 2020; Schrimpf et al., 2021). In these studies, the partial alignability of brain and model representations is interpreted as evidence that brain and models might process language in the same way (Caucheteux and King, 2022).", "text": "Studies have looked at the alignability of neural language representations and human brain activations, with more promising results as language models grow better at modeling language (Sassenhagen and Fiebach, 2020; Schrimpf et al., 2021). In these studies, the partial alignability of brain and model representations is interpreted as evidence that brain and models might process language in the same way (Caucheteux and King, 2022)."} | null |
|
3da74941-87ce-468b-8e8a-cd69a132d125 | 2302.06555v2.pdf | text | Cross-modal alignment. The idea of crossmodal retrieval is not new (Lazaridou et al., 2014), but previously it has mostly been studied with practical considerations in mind. Recently, Merullo et al. (2023) showed that language representations in LMs are functionally similar to image representations in VMs, in that a linear transformation applied to an image representation can be used to prompt a language model into producing a relevant caption. We dial back from function and study whether the concept representations converge toward structural similarity (isomorphism). The key question we address is whether despite the lack of explicit grounding, the representations learned by large pretrained language models structurally resemble properties of the physical world as captured by vision models. More related to our work, | null | 442 | 457 | 72 | image/png | 3 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/23", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 2, "bbox": {"l": 71.18010711669922, "t": 304.51641845703125, "r": 292.0832214355469, "b": 76.014892578125, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 828]}], "orig": "Cross-modal alignment. The idea of crossmodal retrieval is not new (Lazaridou et al., 2014), but previously it has mostly been studied with practical considerations in mind. Recently, Merullo et al. (2023) showed that language representations in LMs are functionally similar to image representations in VMs, in that a linear transformation applied to an image representation can be used to prompt a language model into producing a relevant caption. We dial back from function and study whether the concept representations converge toward structural similarity (isomorphism). The key question we address is whether despite the lack of explicit grounding, the representations learned by large pretrained language models structurally resemble properties of the physical world as captured by vision models. More related to our work,", "text": "Cross-modal alignment. The idea of crossmodal retrieval is not new (Lazaridou et al., 2014), but previously it has mostly been studied with practical considerations in mind. Recently, Merullo et al. (2023) showed that language representations in LMs are functionally similar to image representations in VMs, in that a linear transformation applied to an image representation can be used to prompt a language model into producing a relevant caption. We dial back from function and study whether the concept representations converge toward structural similarity (isomorphism). The key question we address is whether despite the lack of explicit grounding, the representations learned by large pretrained language models structurally resemble properties of the physical world as captured by vision models. More related to our work,"} | null |
|
6d136f70-fd12-4469-8bbf-ac08f34149d3 | 2302.06555v2.pdf | picture | null | null | 125 | 108 | 72 | image/png | 3 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/pictures/0", "parent": {"cref": "#/body"}, "children": [], "label": "picture", "prov": [{"page_no": 2, "bbox": {"l": 313.8261413574219, "t": 763.364501953125, "r": 376.35577392578125, "b": 709.6014404296875, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 0]}], "captions": [], "references": [], "footnotes": [], "image": {"mimetype": "image/png", "dpi": 144, "size": {"width": 125.0, "height": 108.0}, "uri": null}, "annotations": []} | null |
|
ca6cf0de-1cb4-44f8-a79b-4cdd409904e2 | 2302.06555v2.pdf | picture | null | null | 124 | 104 | 72 | image/png | 3 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/pictures/1", "parent": {"cref": "#/body"}, "children": [], "label": "picture", "prov": [{"page_no": 2, "bbox": {"l": 313.7725524902344, "t": 701.4967041015625, "r": 376.127685546875, "b": 649.528076171875, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 0]}], "captions": [], "references": [], "footnotes": [], "image": {"mimetype": "image/png", "dpi": 144, "size": {"width": 124.0, "height": 104.0}, "uri": null}, "annotations": []} | null |
|
f1d01f98-a855-444d-84e0-3eae0a49e3fc | 2302.06555v2.pdf | picture | null | null | 123 | 103 | 72 | image/png | 3 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/pictures/2", "parent": {"cref": "#/body"}, "children": [], "label": "picture", "prov": [{"page_no": 2, "bbox": {"l": 313.82672119140625, "t": 640.9229736328125, "r": 375.7113952636719, "b": 589.3406372070312, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 0]}], "captions": [], "references": [], "footnotes": [], "image": {"mimetype": "image/png", "dpi": 144, "size": {"width": 123.0, "height": 103.0}, "uri": null}, "annotations": []} | null |
|
254effb8-5802-4b97-844a-f2e7488e6665 | 2302.06555v2.pdf | picture | null | null | 123 | 104 | 72 | image/png | 3 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/pictures/3", "parent": {"cref": "#/body"}, "children": [], "label": "picture", "prov": [{"page_no": 2, "bbox": {"l": 314.2514343261719, "t": 580.6930541992188, "r": 375.9010314941406, "b": 528.97998046875, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 0]}], "captions": [], "references": [], "footnotes": [], "image": {"mimetype": "image/png", "dpi": 144, "size": {"width": 123.0, "height": 104.0}, "uri": null}, "annotations": []} | null |
|
0cb7a702-d938-4fc6-b040-3cf8abb2bfb4 | 2302.06555v2.pdf | picture | null | null | 124 | 104 | 72 | image/png | 3 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/pictures/4", "parent": {"cref": "#/body"}, "children": [], "label": "picture", "prov": [{"page_no": 2, "bbox": {"l": 377.7059326171875, "t": 762.4067993164062, "r": 439.4959716796875, "b": 710.218505859375, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 0]}], "captions": [], "references": [], "footnotes": [], "image": {"mimetype": "image/png", "dpi": 144, "size": {"width": 124.0, "height": 104.0}, "uri": null}, "annotations": []} | null |
|
655a5b29-665d-4321-a868-f1be33b24d37 | 2302.06555v2.pdf | picture | null | null | 124 | 101 | 72 | image/png | 3 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/pictures/5", "parent": {"cref": "#/body"}, "children": [], "label": "picture", "prov": [{"page_no": 2, "bbox": {"l": 377.6625061035156, "t": 700.3104248046875, "r": 439.6013488769531, "b": 649.751953125, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 0]}], "captions": [], "references": [], "footnotes": [], "image": {"mimetype": "image/png", "dpi": 144, "size": {"width": 124.0, "height": 101.0}, "uri": null}, "annotations": []} | null |
|
9e733fb5-c421-4396-9509-261f80adbc70 | 2302.06555v2.pdf | picture | null | null | 123 | 102 | 72 | image/png | 3 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/pictures/6", "parent": {"cref": "#/body"}, "children": [], "label": "picture", "prov": [{"page_no": 2, "bbox": {"l": 377.5209045410156, "t": 640.98388671875, "r": 439.1893310546875, "b": 589.8641357421875, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 0]}], "captions": [], "references": [], "footnotes": [], "image": {"mimetype": "image/png", "dpi": 144, "size": {"width": 123.0, "height": 102.0}, "uri": null}, "annotations": []} | null |
|
6413aaf0-41b3-4717-8799-d013d7a0fea6 | 2302.06555v2.pdf | caption | Figure 1: Mapping from MAE$_{Huge}$ (images) to OPT$_{30B}$ (text). Gold labels are in green. | null | 438 | 48 | 72 | image/png | 3 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/24", "parent": {"cref": "#/body"}, "children": [], "label": "caption", "prov": [{"page_no": 2, "bbox": {"l": 306.5188903808594, "t": 511.5455017089844, "r": 525.5476684570312, "b": 487.3186340332031, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 93]}], "orig": "Figure 1: Mapping from MAE$_{Huge}$ (images) to OPT$_{30B}$ (text). Gold labels are in green.", "text": "Figure 1: Mapping from MAE$_{Huge}$ (images) to OPT$_{30B}$ (text). Gold labels are in green."} | null |
|
ab01c853-c183-4c88-968b-f0ded2001a60 | 2302.06555v2.pdf | picture | null | Figure 1: Mapping from MAE$_{Huge}$ (images) to OPT$_{30B}$ (text). Gold labels are in green. | 124 | 103 | 72 | image/png | 3 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/pictures/7", "parent": {"cref": "#/body"}, "children": [], "label": "picture", "prov": [{"page_no": 2, "bbox": {"l": 377.6405029296875, "t": 580.822265625, "r": 439.61199951171875, "b": 529.2573852539062, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 93]}], "captions": [{"cref": "#/texts/24"}], "references": [], "footnotes": [], "image": {"mimetype": "image/png", "dpi": 144, "size": {"width": 124.0, "height": 103.0}, "uri": null}, "annotations": []} | null |
|
005ca179-8bd9-467e-a2cb-95f325429644 | 2302.06555v2.pdf | text | Huh et al. (2024) proposes a similar hypothesis, although studying it from a different perspective, and our findings corroborate theirs. | null | 441 | 76 | 72 | image/png | 3 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/25", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 2, "bbox": {"l": 306.4500427246094, "t": 463.1133728027344, "r": 526.9063720703125, "b": 424.90692138671875, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 136]}], "orig": "Huh et al. (2024) proposes a similar hypothesis, although studying it from a different perspective, and our findings corroborate theirs.", "text": "Huh et al. (2024) proposes a similar hypothesis, although studying it from a different perspective, and our findings corroborate theirs."} | null |
|
727ec5f5-36bf-4d02-b9c4-e4782346cbcf | 2302.06555v2.pdf | section_header | 3 Methodology | null | 172 | 24 | 72 | image/png | 3 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/26", "parent": {"cref": "#/body"}, "children": [], "label": "section_header", "prov": [{"page_no": 2, "bbox": {"l": 306.342529296875, "t": 412.8549499511719, "r": 392.2894287109375, "b": 400.87078857421875, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 13]}], "orig": "3 Methodology", "text": "3 Methodology", "level": 1} | null |
|
be759c2b-2025-4928-8f5d-db721824f5ac | 2302.06555v2.pdf | text | Our primary objective is to compare the representations derived from VMs and LMs and assess their alignability, i.e. the extent to which LMs converge toward VMs' geometries. In the following sections, we introduce the procedures for obtaining the representations and aligning them, with an illustration of our methodology provided in Figure 2. | null | 442 | 186 | 72 | image/png | 3 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/27", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 2, "bbox": {"l": 306.3543395996094, "t": 390.9617919921875, "r": 527.3591918945312, "b": 297.9390869140625, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 343]}], "orig": "Our primary objective is to compare the representations derived from VMs and LMs and assess their alignability, i.e. the extent to which LMs converge toward VMs' geometries. In the following sections, we introduce the procedures for obtaining the representations and aligning them, with an illustration of our methodology provided in Figure 2.", "text": "Our primary objective is to compare the representations derived from VMs and LMs and assess their alignability, i.e. the extent to which LMs converge toward VMs' geometries. In the following sections, we introduce the procedures for obtaining the representations and aligning them, with an illustration of our methodology provided in Figure 2."} | null |
|
7a40986a-6ccd-463e-88ae-91d4adc8cb14 | 2302.06555v2.pdf | text | Vision models. We include fourteen VMs in our experiments, representing three model families: SegFormer (Xie et al., 2021), MAE (He et al., 2022), and ResNet (He et al., 2016). For all three types of VMs, we only employ the encoder component as a visual feature extractor. 4 | null | 442 | 157 | 72 | image/png | 3 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/28", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 2, "bbox": {"l": 306.2715148925781, "t": 288.335693359375, "r": 527.359375, "b": 210.1456298828125, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 274]}], "orig": "Vision models. We include fourteen VMs in our experiments, representing three model families: SegFormer (Xie et al., 2021), MAE (He et al., 2022), and ResNet (He et al., 2016). For all three types of VMs, we only employ the encoder component as a visual feature extractor. 4", "text": "Vision models. We include fourteen VMs in our experiments, representing three model families: SegFormer (Xie et al., 2021), MAE (He et al., 2022), and ResNet (He et al., 2016). For all three types of VMs, we only employ the encoder component as a visual feature extractor. 4"} | null |
|
2d6f6757-f889-4c4f-84e5-e1b6fca83920 | 2302.06555v2.pdf | text | SegFormer models consist of a Transformerbased encoder and a light-weight feed-forward decoder. They are pretrained on object classification data and finetuned on scene parsing data for scene segmentation and object classification. We hypothesize that the reasoning necessary to | null | 443 | 158 | 72 | image/png | 3 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/29", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 2, "bbox": {"l": 306.04150390625, "t": 207.15533447265625, "r": 527.4514770507812, "b": 127.76580810546875, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 278]}], "orig": "SegFormer models consist of a Transformerbased encoder and a light-weight feed-forward decoder. They are pretrained on object classification data and finetuned on scene parsing data for scene segmentation and object classification. We hypothesize that the reasoning necessary to", "text": "SegFormer models consist of a Transformerbased encoder and a light-weight feed-forward decoder. They are pretrained on object classification data and finetuned on scene parsing data for scene segmentation and object classification. We hypothesize that the reasoning necessary to"} | null |
|
8572076e-cbfc-4318-8d1d-4c83905882ef | 2302.06555v2.pdf | footnote | $^{4}$We ran experiments with CLIP (Radford et al., 2021), but report on these separately, since CLIP does not meet the criteria of our study, being trained on a mixture of text and images. CLIP results are presented in Appendix C. | null | 440 | 86 | 72 | image/png | 3 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/30", "parent": {"cref": "#/body"}, "children": [], "label": "footnote", "prov": [{"page_no": 2, "bbox": {"l": 306.3114013671875, "t": 119.45751953125, "r": 526.66015625, "b": 76.37322998046875, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 231]}], "orig": "$^{4}$We ran experiments with CLIP (Radford et al., 2021), but report on these separately, since CLIP does not meet the criteria of our study, being trained on a mixture of text and images. CLIP results are presented in Appendix C.", "text": "$^{4}$We ran experiments with CLIP (Radford et al., 2021), but report on these separately, since CLIP does not meet the criteria of our study, being trained on a mixture of text and images. CLIP results are presented in Appendix C."} | null |
|
2ac5a30c-715e-483c-9d3d-113e3844e6a6 | 2302.06555v2.pdf | caption | Figure 2: Experiments stages: During our experiments, words, sentences, and images are selected from the aliases list (wordlist and ImageNet-21K aliases), Wikipedia and ImageNet-21K, respectively. The source and target spaces are constructed utilizing image and word embeddings which are extracted by specialized vision and language models. | null | 909 | 103 | 72 | image/png | 4 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/31", "parent": {"cref": "#/body"}, "children": [], "label": "caption", "prov": [{"page_no": 3, "bbox": {"l": 71.41864013671875, "t": 595.5836181640625, "r": 525.9263305664062, "b": 543.7525024414062, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 340]}], "orig": "Figure 2: Experiments stages: During our experiments, words, sentences, and images are selected from the aliases list (wordlist and ImageNet-21K aliases), Wikipedia and ImageNet-21K, respectively. The source and target spaces are constructed utilizing image and word embeddings which are extracted by specialized vision and language models.", "text": "Figure 2: Experiments stages: During our experiments, words, sentences, and images are selected from the aliases list (wordlist and ImageNet-21K aliases), Wikipedia and ImageNet-21K, respectively. The source and target spaces are constructed utilizing image and word embeddings which are extracted by specialized vision and language models."} | null |
|
7d1defea-4674-4d83-93b4-7dc5fd44cef9 | 2302.06555v2.pdf | picture | null | Figure 2: Experiments stages: During our experiments, words, sentences, and images are selected from the aliases list (wordlist and ImageNet-21K aliases), Wikipedia and ImageNet-21K, respectively. The source and target spaces are constructed utilizing image and word embeddings which are extracted by specialized vision and language models. | 909 | 332 | 72 | image/png | 4 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/pictures/8", "parent": {"cref": "#/body"}, "children": [], "label": "picture", "prov": [{"page_no": 3, "bbox": {"l": 70.01079559326172, "t": 777.8876953125, "r": 524.3985595703125, "b": 611.6873168945312, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 340]}], "captions": [{"cref": "#/texts/31"}], "references": [], "footnotes": [], "image": {"mimetype": "image/png", "dpi": 144, "size": {"width": 909.0, "height": 332.0}, "uri": null}, "annotations": []} | null |
|
00423384-00ba-411c-87d9-5681643567ad | 2302.06555v2.pdf | text | perform segmentation in context promotes representations that are more similar to those of LMs, which also operate in a discrete space (a vocabulary). The SegFormer models we use are pretrained with ImageNet-1K (Russakovsky et al., 2015) and finetuned with ADE20K (Zhou et al., 2017). | null | 442 | 157 | 72 | image/png | 4 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/32", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 3, "bbox": {"l": 71.2052993774414, "t": 519.85791015625, "r": 292.0829162597656, "b": 441.4917297363281, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 284]}], "orig": "perform segmentation in context promotes representations that are more similar to those of LMs, which also operate in a discrete space (a vocabulary). The SegFormer models we use are pretrained with ImageNet-1K (Russakovsky et al., 2015) and finetuned with ADE20K (Zhou et al., 2017).", "text": "perform segmentation in context promotes representations that are more similar to those of LMs, which also operate in a discrete space (a vocabulary). The SegFormer models we use are pretrained with ImageNet-1K (Russakovsky et al., 2015) and finetuned with ADE20K (Zhou et al., 2017)."} | null |
|
d85c634d-dbb3-48ac-9c94-342db0c51fb3 | 2302.06555v2.pdf | text | MAE models relies on a Transformer-based encoder-decoder architecture, with the VisionTransformer (ViT) (Dosovitskiy et al., 2021) as the encoder backbone. MAE models are trained to reconstruct masked patches in images, i.e., a fully unsupervised training objective, similar to masked language modeling. The encoder takes as input the unmasked image patches, while a lightweight decoder reconstructs the original image from the latent representation of unmasked patches interleaved with mask tokens. The MAE models we use are pretrained on ImageNet-1K. | null | 442 | 320 | 72 | image/png | 4 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/33", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 3, "bbox": {"l": 71.08782196044922, "t": 437.7105407714844, "r": 292.07537841796875, "b": 277.903564453125, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 552]}], "orig": "MAE models relies on a Transformer-based encoder-decoder architecture, with the VisionTransformer (ViT) (Dosovitskiy et al., 2021) as the encoder backbone. MAE models are trained to reconstruct masked patches in images, i.e., a fully unsupervised training objective, similar to masked language modeling. The encoder takes as input the unmasked image patches, while a lightweight decoder reconstructs the original image from the latent representation of unmasked patches interleaved with mask tokens. The MAE models we use are pretrained on ImageNet-1K.", "text": "MAE models relies on a Transformer-based encoder-decoder architecture, with the VisionTransformer (ViT) (Dosovitskiy et al., 2021) as the encoder backbone. MAE models are trained to reconstruct masked patches in images, i.e., a fully unsupervised training objective, similar to masked language modeling. The encoder takes as input the unmasked image patches, while a lightweight decoder reconstructs the original image from the latent representation of unmasked patches interleaved with mask tokens. The MAE models we use are pretrained on ImageNet-1K."} | null |
|
a5381d8b-f763-484f-9b97-4bb4c9d89391 | 2302.06555v2.pdf | text | ResNet models for object classification consist of a bottleneck convolutional neural network with residual blocks as an encoder, with a classification head. They are pretrained on the ImageNet-1K. | null | 438 | 104 | 72 | image/png | 4 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/34", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 3, "bbox": {"l": 71.25859069824219, "t": 274.5047607421875, "r": 290.271728515625, "b": 222.460205078125, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 196]}], "orig": "ResNet models for object classification consist of a bottleneck convolutional neural network with residual blocks as an encoder, with a classification head. They are pretrained on the ImageNet-1K.", "text": "ResNet models for object classification consist of a bottleneck convolutional neural network with residual blocks as an encoder, with a classification head. They are pretrained on the ImageNet-1K."} | null |
|
5adb76e8-32a6-481f-924e-e903b49075e4 | 2302.06555v2.pdf | text | Language models. We include fourteen Transformer-based LMs in our experiments, representing four model families: BERT (Devlin et al., 2019), GPT-2 (Radford et al., 2019), OPT (Zhang et al., 2022) and LLaMA-2 (Touvron et al., 2023). We use six different sizes of BERT (all uncased): BERT$_{Base}$ and BERT$_{Large}$, which are pretrained on the BooksCorpus (Zhu et al., 2015) and English Wikipedia (Foundation), and four smaller BERT sizes, distilled from BERT$_{Large}$ (Turc et al., 2019). | null | 443 | 266 | 72 | image/png | 4 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/35", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 3, "bbox": {"l": 70.57910919189453, "t": 208.846435546875, "r": 292.1816101074219, "b": 76.1505126953125, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 490]}], "orig": "Language models. We include fourteen Transformer-based LMs in our experiments, representing four model families: BERT (Devlin et al., 2019), GPT-2 (Radford et al., 2019), OPT (Zhang et al., 2022) and LLaMA-2 (Touvron et al., 2023). We use six different sizes of BERT (all uncased): BERT$_{Base}$ and BERT$_{Large}$, which are pretrained on the BooksCorpus (Zhu et al., 2015) and English Wikipedia (Foundation), and four smaller BERT sizes, distilled from BERT$_{Large}$ (Turc et al., 2019).", "text": "Language models. We include fourteen Transformer-based LMs in our experiments, representing four model families: BERT (Devlin et al., 2019), GPT-2 (Radford et al., 2019), OPT (Zhang et al., 2022) and LLaMA-2 (Touvron et al., 2023). We use six different sizes of BERT (all uncased): BERT$_{Base}$ and BERT$_{Large}$, which are pretrained on the BooksCorpus (Zhu et al., 2015) and English Wikipedia (Foundation), and four smaller BERT sizes, distilled from BERT$_{Large}$ (Turc et al., 2019)."} | null |
|
35c1f1fe-1b43-4ff4-8be7-b962ec994878 | 2302.06555v2.pdf | text | GPT-2, an auto-regressive decoder-only LM, comes in three sizes, pretrained on the WebText dataset (Radford et al., 2019). OPT also comes in three sizes, pretrained on the union of five datasets (Zhang et al., 2022). LLaMA-2 was pretrained on two trillion tokens. | null | 441 | 158 | 72 | image/png | 4 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/36", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 3, "bbox": {"l": 306.2762145996094, "t": 520.4324340820312, "r": 526.9061889648438, "b": 441.523681640625, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 263]}], "orig": "GPT-2, an auto-regressive decoder-only LM, comes in three sizes, pretrained on the WebText dataset (Radford et al., 2019). OPT also comes in three sizes, pretrained on the union of five datasets (Zhang et al., 2022). LLaMA-2 was pretrained on two trillion tokens.", "text": "GPT-2, an auto-regressive decoder-only LM, comes in three sizes, pretrained on the WebText dataset (Radford et al., 2019). OPT also comes in three sizes, pretrained on the union of five datasets (Zhang et al., 2022). LLaMA-2 was pretrained on two trillion tokens."} | null |
|
1d1c5a43-a827-4da9-af56-3b13c0e76332 | 2302.06555v2.pdf | text | Vision representations. The visual representation of a concept is obtained by embedding the images available for the concept with a given VM encoder and then averaging these representations. When applying SegFormer, we average the patches' representations from the last hidden state as the basis for every image, whereas we use the penultimate hidden state for MAE models. 5 ResNet models generate a single vector per input image from the average pooling layer. | null | 442 | 266 | 72 | image/png | 4 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/37", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 3, "bbox": {"l": 306.2260437011719, "t": 429.16314697265625, "r": 527.4528198242188, "b": 296.0677490234375, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 461]}], "orig": "Vision representations. The visual representation of a concept is obtained by embedding the images available for the concept with a given VM encoder and then averaging these representations. When applying SegFormer, we average the patches' representations from the last hidden state as the basis for every image, whereas we use the penultimate hidden state for MAE models. 5 ResNet models generate a single vector per input image from the average pooling layer.", "text": "Vision representations. The visual representation of a concept is obtained by embedding the images available for the concept with a given VM encoder and then averaging these representations. When applying SegFormer, we average the patches' representations from the last hidden state as the basis for every image, whereas we use the penultimate hidden state for MAE models. 5 ResNet models generate a single vector per input image from the average pooling layer."} | null |
|
a561ca2b-6f29-40bb-88cd-8a7475f2c7d2 | 2302.06555v2.pdf | text | Language representations. The LMs included here were trained on text segments, so applying them to words in isolation could result in unpredictable behavior. We therefore represent words by embedding English Wikipedia sentences, using the token representations that form the concept, decontextualizing these representations by averaging across different sentences (Abdou et al., 2021). In the case of masked language models, we employ | null | 442 | 239 | 72 | image/png | 4 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/38", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 3, "bbox": {"l": 306.23394775390625, "t": 283.43695068359375, "r": 527.3591918945312, "b": 164.11749267578125, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 434]}], "orig": "Language representations. The LMs included here were trained on text segments, so applying them to words in isolation could result in unpredictable behavior. We therefore represent words by embedding English Wikipedia sentences, using the token representations that form the concept, decontextualizing these representations by averaging across different sentences (Abdou et al., 2021). In the case of masked language models, we employ", "text": "Language representations. The LMs included here were trained on text segments, so applying them to words in isolation could result in unpredictable behavior. We therefore represent words by embedding English Wikipedia sentences, using the token representations that form the concept, decontextualizing these representations by averaging across different sentences (Abdou et al., 2021). In the case of masked language models, we employ"} | null |
|
58b5272c-7ffa-4e31-a994-75ac7501824f | 2302.06555v2.pdf | footnote | $^{5}$We also experimented with utilizing the representations from the last hidden state; however, the results were not as promising as those obtained from the penultimate hidden state. Caron et al. (2021) demonstrate the penultimate-layer features in ViTs trained with DINO exhibit strong correlations with saliency information in the visual input, such as object boundaries and so on. | null | 441 | 151 | 72 | image/png | 4 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/39", "parent": {"cref": "#/body"}, "children": [], "label": "footnote", "prov": [{"page_no": 3, "bbox": {"l": 306.2687072753906, "t": 152.629638671875, "r": 527.111083984375, "b": 76.98725891113281, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 386]}], "orig": "$^{5}$We also experimented with utilizing the representations from the last hidden state; however, the results were not as promising as those obtained from the penultimate hidden state. Caron et al. (2021) demonstrate the penultimate-layer features in ViTs trained with DINO exhibit strong correlations with saliency information in the visual input, such as object boundaries and so on.", "text": "$^{5}$We also experimented with utilizing the representations from the last hidden state; however, the results were not as promising as those obtained from the penultimate hidden state. Caron et al. (2021) demonstrate the penultimate-layer features in ViTs trained with DINO exhibit strong correlations with saliency information in the visual input, such as object boundaries and so on."} | null |
|
1bb9027b-a593-4c15-a20d-54e53fbb922d | 2302.06555v2.pdf | text | an averaging approach on the token representations forming the concept; otherwise, we choose for the last token within the concept (Zou et al., 2023). | null | 438 | 75 | 72 | image/png | 5 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/40", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 4, "bbox": {"l": 71.54341125488281, "t": 776.0916748046875, "r": 290.26824951171875, "b": 738.40966796875, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 150]}], "orig": "an averaging approach on the token representations forming the concept; otherwise, we choose for the last token within the concept (Zou et al., 2023).", "text": "an averaging approach on the token representations forming the concept; otherwise, we choose for the last token within the concept (Zou et al., 2023)."} | null |
|
4e3bb844-dfc1-4d69-8906-303f9098040b | 2302.06555v2.pdf | text | Linear projection. Since we are interested in the extent to which vision and language representations are isomorphic, we focus on linear projections. 6 Following Conneau et al. (2018), we use Procrustes analysis (Schönemann, 1966) to align the representations of VMs to those of LMs, given a bimodal dictionary (§ 4.1). Given the VM matrix A (i.e., the visual representations of concepts) and the LM matrix B (i.e. the language representation of the concepts) we use Procrustes analysis to find the orthogonal matrix Ω that most closely maps source space A onto the target space B . Given the constrain of orthogonality the optimization Ω = min$_{R}$ ∥ RA - B ∥$_{F}$ , s.t. R $^{T}$R = I has the closed form solution Ω = UV $^{T}$,U Σ V = SVD ( BA $^{T}$) , where SVD stands for singular value decomposition. We induce the alignment from a small set of dictionary pairs, evaluating it on held-out data (§ 4.2). Given the necessity for both the source and target space to have the same dimensionality, we employ principal component analysis (PCA) to reduce the dimensionality of the larger space in cases of a mismatch. 7 | null | 442 | 618 | 72 | image/png | 5 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/41", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 4, "bbox": {"l": 70.89579772949219, "t": 726.5772705078125, "r": 292.08294677734375, "b": 417.6516418457031, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 1121]}], "orig": "Linear projection. Since we are interested in the extent to which vision and language representations are isomorphic, we focus on linear projections. 6 Following Conneau et al. (2018), we use Procrustes analysis (Sch\u00f6nemann, 1966) to align the representations of VMs to those of LMs, given a bimodal dictionary (\u00a7 4.1). Given the VM matrix A (i.e., the visual representations of concepts) and the LM matrix B (i.e. the language representation of the concepts) we use Procrustes analysis to find the orthogonal matrix \u2126 that most closely maps source space A onto the target space B . Given the constrain of orthogonality the optimization \u2126 = min$_{R}$ \u2225 RA - B \u2225$_{F}$ , s.t. R $^{T}$R = I has the closed form solution \u2126 = UV $^{T}$,U \u03a3 V = SVD ( BA $^{T}$) , where SVD stands for singular value decomposition. We induce the alignment from a small set of dictionary pairs, evaluating it on held-out data (\u00a7 4.2). Given the necessity for both the source and target space to have the same dimensionality, we employ principal component analysis (PCA) to reduce the dimensionality of the larger space in cases of a mismatch. 7", "text": "Linear projection. Since we are interested in the extent to which vision and language representations are isomorphic, we focus on linear projections. 6 Following Conneau et al. (2018), we use Procrustes analysis (Sch\u00f6nemann, 1966) to align the representations of VMs to those of LMs, given a bimodal dictionary (\u00a7 4.1). Given the VM matrix A (i.e., the visual representations of concepts) and the LM matrix B (i.e. the language representation of the concepts) we use Procrustes analysis to find the orthogonal matrix \u2126 that most closely maps source space A onto the target space B . Given the constrain of orthogonality the optimization \u2126 = min$_{R}$ \u2225 RA - B \u2225$_{F}$ , s.t. R $^{T}$R = I has the closed form solution \u2126 = UV $^{T}$,U \u03a3 V = SVD ( BA $^{T}$) , where SVD stands for singular value decomposition. We induce the alignment from a small set of dictionary pairs, evaluating it on held-out data (\u00a7 4.2). Given the necessity for both the source and target space to have the same dimensionality, we employ principal component analysis (PCA) to reduce the dimensionality of the larger space in cases of a mismatch. 7"} | null |
|
db03dd4e-1bdf-433d-b4bd-05021749e207 | 2302.06555v2.pdf | section_header | 4 Experimental Setup | null | 242 | 24 | 72 | image/png | 5 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/42", "parent": {"cref": "#/body"}, "children": [], "label": "section_header", "prov": [{"page_no": 4, "bbox": {"l": 71.08319091796875, "t": 403.1234130859375, "r": 191.88674926757812, "b": 390.90045166015625, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 20]}], "orig": "4 Experimental Setup", "text": "4 Experimental Setup", "level": 1} | null |
|
33c633e8-b91c-4fde-90cf-a16b59f21946 | 2302.06555v2.pdf | text | In this section, we discuss details around bimodal dictionary compilation (§ 4.1), evaluation metrics, as well as our baselines (§ 4.2). | null | 441 | 76 | 72 | image/png | 5 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/43", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 4, "bbox": {"l": 71.15853881835938, "t": 380.3851623535156, "r": 291.63519287109375, "b": 342.40765380859375, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 136]}], "orig": "In this section, we discuss details around bimodal dictionary compilation (\u00a7 4.1), evaluation metrics, as well as our baselines (\u00a7 4.2).", "text": "In this section, we discuss details around bimodal dictionary compilation (\u00a7 4.1), evaluation metrics, as well as our baselines (\u00a7 4.2)."} | null |
|
288e8dfc-c707-4f66-b71e-f47ce763588a | 2302.06555v2.pdf | section_header | 4.1 Bimodal Dictionary Compilation | null | 357 | 22 | 72 | image/png | 5 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/44", "parent": {"cref": "#/body"}, "children": [], "label": "section_header", "prov": [{"page_no": 4, "bbox": {"l": 71.00336456298828, "t": 329.13519287109375, "r": 249.28378295898438, "b": 317.93157958984375, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 34]}], "orig": "4.1 Bimodal Dictionary Compilation", "text": "4.1 Bimodal Dictionary Compilation", "level": 1} | null |
|
1afe2f60-0a37-4fee-b2e2-ebdd5f6eb3c1 | 2302.06555v2.pdf | text | We build bimodal dictionaries of image-text pairs based on the ImageNet21K dataset (Russakovsky et al., 2015) and the CLDI (cross-lingual dictionary induction) dataset (Hartmann and Søgaard, 2018). In ImageNet, a concept class has a unique ID and is represented by multiple images and one or more names (which we refer to as aliases ), many of which are multi-word expressions. We filter the data from ImageNet-21K: keeping classes with over 100 images available, aliases that appear at least five times in Wikipedia, and classes with at | null | 442 | 295 | 72 | image/png | 5 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/45", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 4, "bbox": {"l": 70.98979187011719, "t": 310.3228759765625, "r": 292.07537841796875, "b": 163.00567626953125, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 537]}], "orig": "We build bimodal dictionaries of image-text pairs based on the ImageNet21K dataset (Russakovsky et al., 2015) and the CLDI (cross-lingual dictionary induction) dataset (Hartmann and S\u00f8gaard, 2018). In ImageNet, a concept class has a unique ID and is represented by multiple images and one or more names (which we refer to as aliases ), many of which are multi-word expressions. We filter the data from ImageNet-21K: keeping classes with over 100 images available, aliases that appear at least five times in Wikipedia, and classes with at", "text": "We build bimodal dictionaries of image-text pairs based on the ImageNet21K dataset (Russakovsky et al., 2015) and the CLDI (cross-lingual dictionary induction) dataset (Hartmann and S\u00f8gaard, 2018). In ImageNet, a concept class has a unique ID and is represented by multiple images and one or more names (which we refer to as aliases ), many of which are multi-word expressions. We filter the data from ImageNet-21K: keeping classes with over 100 images available, aliases that appear at least five times in Wikipedia, and classes with at"} | null |
|
c20d0945-d3c3-440c-9645-e23e6f264b44 | 2302.06555v2.pdf | footnote | $^{6}$For work on non-linear projection between representation spaces, see Nakashole (2018); Zhao and Gilman (2020); Glavaš and Vuli´c (2020). | null | 441 | 63 | 72 | image/png | 5 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/46", "parent": {"cref": "#/body"}, "children": [], "label": "footnote", "prov": [{"page_no": 4, "bbox": {"l": 71.50354766845703, "t": 152.3046875, "r": 291.7594909667969, "b": 120.85226440429688, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 142]}], "orig": "$^{6}$For work on non-linear projection between representation spaces, see Nakashole (2018); Zhao and Gilman (2020); Glava\u0161 and Vuli\u00b4c (2020).", "text": "$^{6}$For work on non-linear projection between representation spaces, see Nakashole (2018); Zhao and Gilman (2020); Glava\u0161 and Vuli\u00b4c (2020)."} | null |
|
8242f6ac-1f27-4b3f-9a93-f2fdc5fe0d78 | 2302.06555v2.pdf | footnote | $^{7}$The variance is retained for most models after dimensionality reduction, except for a few cases where there is some loss of information. The cumulative of explained variance ratios for different models are presented in Table 8. | null | 441 | 85 | 72 | image/png | 5 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/47", "parent": {"cref": "#/body"}, "children": [], "label": "footnote", "prov": [{"page_no": 4, "bbox": {"l": 71.3525619506836, "t": 118.89080810546875, "r": 291.7541809082031, "b": 76.62451171875, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 233]}], "orig": "$^{7}$The variance is retained for most models after dimensionality reduction, except for a few cases where there is some loss of information. The cumulative of explained variance ratios for different models are presented in Table 8.", "text": "$^{7}$The variance is retained for most models after dimensionality reduction, except for a few cases where there is some loss of information. The cumulative of explained variance ratios for different models are presented in Table 8."} | null |
|
65f280ac-2ac5-43b8-903c-d0c2b8f47d32 | 2302.06555v2.pdf | caption | Table 1: Statistics of the bimodal dictionaries. | null | 408 | 21 | 72 | image/png | 5 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/48", "parent": {"cref": "#/body"}, "children": [], "label": "caption", "prov": [{"page_no": 4, "bbox": {"l": 313.8568420410156, "t": 716.146240234375, "r": 517.947021484375, "b": 705.5986328125, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 48]}], "orig": "Table 1: Statistics of the bimodal dictionaries.", "text": "Table 1: Statistics of the bimodal dictionaries."} | null |
|
958fcef6-7205-483c-a140-0bc9acb055f1 | 2302.06555v2.pdf | table | <table><tbody><tr><th>Set</th><th>Num. of classes</th><th>Num. of aliases</th><th>Num. of pairs</th></tr><tr><td>Only-1K</td><td>491</td><td>655</td><td>655</td></tr><tr><td>Exclude-1K</td><td>5,942</td><td>7,194</td><td>7,194</td></tr><tr><td>EN-CLDI</td><td>1,690</td><td>1,690</td><td>1,690</td></tr></tbody></table> | Table 1: Statistics of the bimodal dictionaries. | 435 | 103 | 72 | image/png | 5 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/tables/0", "parent": {"cref": "#/body"}, "children": [], "label": "table", "prov": [{"page_no": 4, "bbox": {"l": 308.0361328125, "t": 778.875, "r": 525.6458740234375, "b": 727.481201171875, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 0]}], "captions": [{"cref": "#/texts/48"}], "references": [], "footnotes": [], "image": {"mimetype": "image/png", "dpi": 144, "size": {"width": 435.0, "height": 103.0}, "uri": null}, "data": {"table_cells": [{"bbox": {"l": 313.4814453125, "t": 774.6961669921875, "r": 323.4200134277344, "b": 767.7437744140625, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 0, "end_row_offset_idx": 1, "start_col_offset_idx": 0, "end_col_offset_idx": 1, "text": "Set", "column_header": true, "row_header": false, "row_section": false}, {"bbox": {"l": 361.808349609375, "t": 774.6961669921875, "r": 411.7344970703125, "b": 767.7437744140625, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 0, "end_row_offset_idx": 1, "start_col_offset_idx": 1, "end_col_offset_idx": 2, "text": "Num. of classes", "column_header": true, "row_header": false, "row_section": false}, {"bbox": {"l": 420.2577209472656, "t": 774.6961669921875, "r": 469.3206787109375, "b": 767.7437744140625, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 0, "end_row_offset_idx": 1, "start_col_offset_idx": 2, "end_col_offset_idx": 3, "text": "Num. of aliases", "column_header": true, "row_header": false, "row_section": false}, {"bbox": {"l": 477.8439025878906, "t": 774.6961669921875, "r": 521.2921142578125, "b": 767.7437744140625, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 0, "end_row_offset_idx": 1, "start_col_offset_idx": 3, "end_col_offset_idx": 4, "text": "Num. of pairs", "column_header": true, "row_header": false, "row_section": false}, {"bbox": {"l": 313.4814453125, "t": 760.9743041992188, "r": 341.1274719238281, "b": 754.02197265625, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 1, "end_row_offset_idx": 2, "start_col_offset_idx": 0, "end_col_offset_idx": 1, "text": "Only-1K", "column_header": false, "row_header": true, "row_section": false}, {"bbox": {"l": 400.0690002441406, "t": 760.9743041992188, "r": 411.7339782714844, "b": 754.02197265625, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 1, "end_row_offset_idx": 2, "start_col_offset_idx": 1, "end_col_offset_idx": 2, "text": "491", "column_header": false, "row_header": false, "row_section": false}, {"bbox": {"l": 457.6551818847656, "t": 760.9743041992188, "r": 469.3201599121094, "b": 754.02197265625, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 1, "end_row_offset_idx": 2, "start_col_offset_idx": 2, "end_col_offset_idx": 3, "text": "655", "column_header": false, "row_header": false, "row_section": false}, {"bbox": {"l": 509.6265869140625, "t": 760.9743041992188, "r": 521.2916259765625, "b": 754.02197265625, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 1, "end_row_offset_idx": 2, "start_col_offset_idx": 3, "end_col_offset_idx": 4, "text": "655", "column_header": false, "row_header": false, "row_section": false}, {"bbox": {"l": 313.4814453125, "t": 751.3157958984375, "r": 351.05828857421875, "b": 744.3634643554688, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 2, "end_row_offset_idx": 3, "start_col_offset_idx": 0, "end_col_offset_idx": 1, "text": "Exclude-1K", "column_header": false, "row_header": true, "row_section": false}, {"bbox": {"l": 394.2363586425781, "t": 751.3157958984375, "r": 411.7338562011719, "b": 744.3634643554688, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 2, "end_row_offset_idx": 3, "start_col_offset_idx": 1, "end_col_offset_idx": 2, "text": "5,942", "column_header": false, "row_header": false, "row_section": false}, {"bbox": {"l": 451.8225402832031, "t": 751.3157958984375, "r": 469.3200378417969, "b": 744.3634643554688, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 2, "end_row_offset_idx": 3, "start_col_offset_idx": 2, "end_col_offset_idx": 3, "text": "7,194", "column_header": false, "row_header": false, "row_section": false}, {"bbox": {"l": 503.7939758300781, "t": 751.3157958984375, "r": 521.2914428710938, "b": 744.3634643554688, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 2, "end_row_offset_idx": 3, "start_col_offset_idx": 3, "end_col_offset_idx": 4, "text": "7,194", "column_header": false, "row_header": false, "row_section": false}, {"bbox": {"l": 313.4814453125, "t": 737.5939331054688, "r": 344.5802917480469, "b": 730.6416015625, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 3, "end_row_offset_idx": 4, "start_col_offset_idx": 0, "end_col_offset_idx": 1, "text": "EN-CLDI", "column_header": false, "row_header": true, "row_section": false}, {"bbox": {"l": 394.2363586425781, "t": 737.5939331054688, "r": 411.7338562011719, "b": 730.6416015625, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 3, "end_row_offset_idx": 4, "start_col_offset_idx": 1, "end_col_offset_idx": 2, "text": "1,690", "column_header": false, "row_header": false, "row_section": false}, {"bbox": {"l": 451.8225402832031, "t": 737.5939331054688, "r": 469.3200378417969, "b": 730.6416015625, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 3, "end_row_offset_idx": 4, "start_col_offset_idx": 2, "end_col_offset_idx": 3, "text": "1,690", "column_header": false, "row_header": false, "row_section": false}, {"bbox": {"l": 503.7939758300781, "t": 737.5939331054688, "r": 521.2914428710938, "b": 730.6416015625, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 3, "end_row_offset_idx": 4, "start_col_offset_idx": 3, "end_col_offset_idx": 4, "text": "1,690", "column_header": false, "row_header": false, "row_section": false}], "num_rows": 4, "num_cols": 4, "grid": [[{"bbox": {"l": 313.4814453125, "t": 774.6961669921875, "r": 323.4200134277344, "b": 767.7437744140625, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 0, "end_row_offset_idx": 1, "start_col_offset_idx": 0, "end_col_offset_idx": 1, "text": "Set", "column_header": true, "row_header": false, "row_section": false}, {"bbox": {"l": 361.808349609375, "t": 774.6961669921875, "r": 411.7344970703125, "b": 767.7437744140625, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 0, "end_row_offset_idx": 1, "start_col_offset_idx": 1, "end_col_offset_idx": 2, "text": "Num. of classes", "column_header": true, "row_header": false, "row_section": false}, {"bbox": {"l": 420.2577209472656, "t": 774.6961669921875, "r": 469.3206787109375, "b": 767.7437744140625, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 0, "end_row_offset_idx": 1, "start_col_offset_idx": 2, "end_col_offset_idx": 3, "text": "Num. of aliases", "column_header": true, "row_header": false, "row_section": false}, {"bbox": {"l": 477.8439025878906, "t": 774.6961669921875, "r": 521.2921142578125, "b": 767.7437744140625, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 0, "end_row_offset_idx": 1, "start_col_offset_idx": 3, "end_col_offset_idx": 4, "text": "Num. of pairs", "column_header": true, "row_header": false, "row_section": false}], [{"bbox": {"l": 313.4814453125, "t": 760.9743041992188, "r": 341.1274719238281, "b": 754.02197265625, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 1, "end_row_offset_idx": 2, "start_col_offset_idx": 0, "end_col_offset_idx": 1, "text": "Only-1K", "column_header": false, "row_header": true, "row_section": false}, {"bbox": {"l": 400.0690002441406, "t": 760.9743041992188, "r": 411.7339782714844, "b": 754.02197265625, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 1, "end_row_offset_idx": 2, "start_col_offset_idx": 1, "end_col_offset_idx": 2, "text": "491", "column_header": false, "row_header": false, "row_section": false}, {"bbox": {"l": 457.6551818847656, "t": 760.9743041992188, "r": 469.3201599121094, "b": 754.02197265625, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 1, "end_row_offset_idx": 2, "start_col_offset_idx": 2, "end_col_offset_idx": 3, "text": "655", "column_header": false, "row_header": false, "row_section": false}, {"bbox": {"l": 509.6265869140625, "t": 760.9743041992188, "r": 521.2916259765625, "b": 754.02197265625, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 1, "end_row_offset_idx": 2, "start_col_offset_idx": 3, "end_col_offset_idx": 4, "text": "655", "column_header": false, "row_header": false, "row_section": false}], [{"bbox": {"l": 313.4814453125, "t": 751.3157958984375, "r": 351.05828857421875, "b": 744.3634643554688, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 2, "end_row_offset_idx": 3, "start_col_offset_idx": 0, "end_col_offset_idx": 1, "text": "Exclude-1K", "column_header": false, "row_header": true, "row_section": false}, {"bbox": {"l": 394.2363586425781, "t": 751.3157958984375, "r": 411.7338562011719, "b": 744.3634643554688, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 2, "end_row_offset_idx": 3, "start_col_offset_idx": 1, "end_col_offset_idx": 2, "text": "5,942", "column_header": false, "row_header": false, "row_section": false}, {"bbox": {"l": 451.8225402832031, "t": 751.3157958984375, "r": 469.3200378417969, "b": 744.3634643554688, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 2, "end_row_offset_idx": 3, "start_col_offset_idx": 2, "end_col_offset_idx": 3, "text": "7,194", "column_header": false, "row_header": false, "row_section": false}, {"bbox": {"l": 503.7939758300781, "t": 751.3157958984375, "r": 521.2914428710938, "b": 744.3634643554688, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 2, "end_row_offset_idx": 3, "start_col_offset_idx": 3, "end_col_offset_idx": 4, "text": "7,194", "column_header": false, "row_header": false, "row_section": false}], [{"bbox": {"l": 313.4814453125, "t": 737.5939331054688, "r": 344.5802917480469, "b": 730.6416015625, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 3, "end_row_offset_idx": 4, "start_col_offset_idx": 0, "end_col_offset_idx": 1, "text": "EN-CLDI", "column_header": false, "row_header": true, "row_section": false}, {"bbox": {"l": 394.2363586425781, "t": 737.5939331054688, "r": 411.7338562011719, "b": 730.6416015625, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 3, "end_row_offset_idx": 4, "start_col_offset_idx": 1, "end_col_offset_idx": 2, "text": "1,690", "column_header": false, "row_header": false, "row_section": false}, {"bbox": {"l": 451.8225402832031, "t": 737.5939331054688, "r": 469.3200378417969, "b": 730.6416015625, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 3, "end_row_offset_idx": 4, "start_col_offset_idx": 2, "end_col_offset_idx": 3, "text": "1,690", "column_header": false, "row_header": false, "row_section": false}, {"bbox": {"l": 503.7939758300781, "t": 737.5939331054688, "r": 521.2914428710938, "b": 730.6416015625, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 3, "end_row_offset_idx": 4, "start_col_offset_idx": 3, "end_col_offset_idx": 4, "text": "1,690", "column_header": false, "row_header": false, "row_section": false}]]}} | null |
|
27d964d3-7aea-42d2-aa69-cddcfe5b13ef | 2302.06555v2.pdf | text | least one alias. As a result, 11,338 classes and 13,460 aliases meet the criteria. We further filter aliases that are shared by two different class IDs, and aliases for which their hyponyms are already in the aliases set. 8 To avoid any form of bias, given that the VMs we experiment with have been pretrained on ImageNet-1K, we report results on ImageNet-21K excluding the concepts in ImageNet-1K (Exclude-1K). | null | 442 | 237 | 72 | image/png | 5 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/49", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 4, "bbox": {"l": 306.3482360839844, "t": 682.125, "r": 527.3558349609375, "b": 563.3076171875, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 411]}], "orig": "least one alias. As a result, 11,338 classes and 13,460 aliases meet the criteria. We further filter aliases that are shared by two different class IDs, and aliases for which their hyponyms are already in the aliases set. 8 To avoid any form of bias, given that the VMs we experiment with have been pretrained on ImageNet-1K, we report results on ImageNet-21K excluding the concepts in ImageNet-1K (Exclude-1K).", "text": "least one alias. As a result, 11,338 classes and 13,460 aliases meet the criteria. We further filter aliases that are shared by two different class IDs, and aliases for which their hyponyms are already in the aliases set. 8 To avoid any form of bias, given that the VMs we experiment with have been pretrained on ImageNet-1K, we report results on ImageNet-21K excluding the concepts in ImageNet-1K (Exclude-1K)."} | null |
|
6a58186a-e7f7-4af5-bed5-9cd344e844e7 | 2302.06555v2.pdf | text | One important limitation of the Exclude-1K bimodal dictionary is that all concepts are nouns. Therefore, to investigate how our results generalize to other parts of speech (POS), we also use the English subset of CLDI dataset (EN-CLDI), which contains images paired with verbs and adjectives. Each word within this set is unique and paired with at least 22 images. Final statistics of the processed datasets are reported in Table 1. | null | 443 | 238 | 72 | image/png | 5 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/50", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 4, "bbox": {"l": 306.14984130859375, "t": 560.3670654296875, "r": 527.4514770507812, "b": 441.3656311035156, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 432]}], "orig": "One important limitation of the Exclude-1K bimodal dictionary is that all concepts are nouns. Therefore, to investigate how our results generalize to other parts of speech (POS), we also use the English subset of CLDI dataset (EN-CLDI), which contains images paired with verbs and adjectives. Each word within this set is unique and paired with at least 22 images. Final statistics of the processed datasets are reported in Table 1.", "text": "One important limitation of the Exclude-1K bimodal dictionary is that all concepts are nouns. Therefore, to investigate how our results generalize to other parts of speech (POS), we also use the English subset of CLDI dataset (EN-CLDI), which contains images paired with verbs and adjectives. Each word within this set is unique and paired with at least 22 images. Final statistics of the processed datasets are reported in Table 1."} | null |
|
4486b15d-794a-4f20-87c0-89454067deb3 | 2302.06555v2.pdf | text | The pairs in these bimodal dictionaries are split 70-30 for training and testing based on the class IDs to avoid train-test leakage. 9 We compute five such splits at random and report averaged results. See § 6 for the impact of training set size variations. | null | 442 | 131 | 72 | image/png | 5 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/51", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 4, "bbox": {"l": 306.2119140625, "t": 438.3727722167969, "r": 527.4554443359375, "b": 373.1012878417969, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 257]}], "orig": "The pairs in these bimodal dictionaries are split 70-30 for training and testing based on the class IDs to avoid train-test leakage. 9 We compute five such splits at random and report averaged results. See \u00a7 6 for the impact of training set size variations.", "text": "The pairs in these bimodal dictionaries are split 70-30 for training and testing based on the class IDs to avoid train-test leakage. 9 We compute five such splits at random and report averaged results. See \u00a7 6 for the impact of training set size variations."} | null |
|
57508efd-4361-4518-8d6f-0a141a1b30a1 | 2302.06555v2.pdf | section_header | 4.2 Evaluation | null | 152 | 22 | 72 | image/png | 5 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/52", "parent": {"cref": "#/body"}, "children": [], "label": "section_header", "prov": [{"page_no": 4, "bbox": {"l": 306.46258544921875, "t": 362.63116455078125, "r": 382.63604736328125, "b": 351.5690002441406, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 14]}], "orig": "4.2 Evaluation", "text": "4.2 Evaluation", "level": 1} | null |
|
fb9d2a8c-5170-4f2d-99f2-e9ede95d7298 | 2302.06555v2.pdf | text | We induce a linear mapping Ω based on training image-text pairs sampled from A and B , respectively. We then evaluate how close A Ω is to B by computing retrieval precision on held-out imagetext pairs. To make the retrieval task as challenging as possible, the target space B is expanded with 65,599 words from an English wordlist in addition to 13,460 aliases, resulting in a total of 79,059 aliases in the final target space. | null | 442 | 240 | 72 | image/png | 5 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/53", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 4, "bbox": {"l": 306.2489013671875, "t": 345.0776672363281, "r": 527.352783203125, "b": 225.149169921875, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 427]}], "orig": "We induce a linear mapping \u2126 based on training image-text pairs sampled from A and B , respectively. We then evaluate how close A \u2126 is to B by computing retrieval precision on held-out imagetext pairs. To make the retrieval task as challenging as possible, the target space B is expanded with 65,599 words from an English wordlist in addition to 13,460 aliases, resulting in a total of 79,059 aliases in the final target space.", "text": "We induce a linear mapping \u2126 based on training image-text pairs sampled from A and B , respectively. We then evaluate how close A \u2126 is to B by computing retrieval precision on held-out imagetext pairs. To make the retrieval task as challenging as possible, the target space B is expanded with 65,599 words from an English wordlist in addition to 13,460 aliases, resulting in a total of 79,059 aliases in the final target space."} | null |
|
e4e28a87-f60f-436a-b661-d73190c39776 | 2302.06555v2.pdf | text | Metrics. We evaluate alignment in terms of precision-atk (P@ k ), a well-established metric employed in the evaluation of multilingual word embeddings (Conneau et al., 2018), with k ∈ { 1 , 10 , 100 } . 10 Note that this performance metric | null | 438 | 134 | 72 | image/png | 5 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/54", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 4, "bbox": {"l": 306.3006286621094, "t": 215.8372802734375, "r": 525.5789184570312, "b": 148.8270263671875, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 239]}], "orig": "Metrics. We evaluate alignment in terms of precision-atk (P@ k ), a well-established metric employed in the evaluation of multilingual word embeddings (Conneau et al., 2018), with k \u2208 { 1 , 10 , 100 } . 10 Note that this performance metric", "text": "Metrics. We evaluate alignment in terms of precision-atk (P@ k ), a well-established metric employed in the evaluation of multilingual word embeddings (Conneau et al., 2018), with k \u2208 { 1 , 10 , 100 } . 10 Note that this performance metric"} | null |
|
510f1161-7607-48cd-8001-42c1c446291e | 2302.06555v2.pdf | footnote | $^{8}$We obtain the aliases hypernyms and hyponyms from the Princeton WordNet (Fellbaum, 2010). | null | 438 | 41 | 72 | image/png | 5 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/55", "parent": {"cref": "#/body"}, "children": [], "label": "footnote", "prov": [{"page_no": 4, "bbox": {"l": 306.5758361816406, "t": 141.65576171875, "r": 525.547119140625, "b": 120.88531494140625, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 95]}], "orig": "$^{8}$We obtain the aliases hypernyms and hyponyms from the Princeton WordNet (Fellbaum, 2010).", "text": "$^{8}$We obtain the aliases hypernyms and hyponyms from the Princeton WordNet (Fellbaum, 2010)."} | null |
|
62693f5e-2a68-4dc6-9215-cfcc0d77a24e | 2302.06555v2.pdf | footnote | $^{9}$In the EN-CLDI set, we simply use words to mitigate the risk of train-test leakage. | null | 437 | 40 | 72 | image/png | 5 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/56", "parent": {"cref": "#/body"}, "children": [], "label": "footnote", "prov": [{"page_no": 4, "bbox": {"l": 306.7635803222656, "t": 119.1673583984375, "r": 525.5420532226562, "b": 98.96624755859375, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 89]}], "orig": "$^{9}$In the EN-CLDI set, we simply use words to mitigate the risk of train-test leakage.", "text": "$^{9}$In the EN-CLDI set, we simply use words to mitigate the risk of train-test leakage."} | null |
|
817261b3-6267-4805-89fa-7af4c9287a4f | 2302.06555v2.pdf | footnote | $^{10}$For example, we could use the mapping of the image of an apple into the word ‘apple’, and the mapping of the image | null | 439 | 42 | 72 | image/png | 5 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/57", "parent": {"cref": "#/body"}, "children": [], "label": "footnote", "prov": [{"page_no": 4, "bbox": {"l": 306.67706298828125, "t": 96.8956298828125, "r": 525.7845458984375, "b": 75.97491455078125, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 121]}], "orig": "$^{10}$For example, we could use the mapping of the image of an apple into the word \u2018apple\u2019, and the mapping of the image", "text": "$^{10}$For example, we could use the mapping of the image of an apple into the word \u2018apple\u2019, and the mapping of the image"} | null |
|
e038a730-a293-4950-a8fd-df077cd03f23 | 2302.06555v2.pdf | caption | Table 2: Alignment results for our baselines. All the Precision@ k scores are reported in percentage. | null | 439 | 49 | 72 | image/png | 6 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/58", "parent": {"cref": "#/body"}, "children": [], "label": "caption", "prov": [{"page_no": 5, "bbox": {"l": 71.00687408447266, "t": 713.8568115234375, "r": 290.2685546875, "b": 689.1964721679688, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 101]}], "orig": "Table 2: Alignment results for our baselines. All the Precision@ k scores are reported in percentage.", "text": "Table 2: Alignment results for our baselines. All the Precision@ k scores are reported in percentage."} | null |
|
908f9f51-c9ec-454a-8c0e-c609f8146b96 | 2302.06555v2.pdf | table | <table><tbody><tr><th>Baseline</th><th>P@1</th><th>P@10</th><th>P@100</th></tr><tr><td>Random retrieval</td><td>0.0015</td><td>0.0153</td><td>0.1531</td></tr><tr><td>Length-frequency alignment</td><td>0.0032</td><td>0.0127</td><td>0.6053</td></tr><tr><td>Non-isomorphic alignment</td><td>0.0000</td><td>0.0121</td><td>0.1105</td></tr></tbody></table> | Table 2: Alignment results for our baselines. All the Precision@ k scores are reported in percentage. | 437 | 111 | 72 | image/png | 6 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/tables/1", "parent": {"cref": "#/body"}, "children": [], "label": "table", "prov": [{"page_no": 5, "bbox": {"l": 72.87548065185547, "t": 779.512451171875, "r": 291.2668762207031, "b": 723.8392333984375, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 0]}], "captions": [{"cref": "#/texts/58"}], "references": [], "footnotes": [], "image": {"mimetype": "image/png", "dpi": 144, "size": {"width": 437.0, "height": 111.0}, "uri": null}, "data": {"table_cells": [{"bbox": {"l": 79.09597778320312, "t": 774.0670776367188, "r": 109.7222900390625, "b": 766.1170654296875, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 0, "end_row_offset_idx": 1, "start_col_offset_idx": 0, "end_col_offset_idx": 1, "text": "Baseline", "column_header": true, "row_header": false, "row_section": false}, {"bbox": {"l": 197.4066619873047, "t": 774.0670776367188, "r": 214.98745727539062, "b": 766.1170654296875, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 0, "end_row_offset_idx": 1, "start_col_offset_idx": 1, "end_col_offset_idx": 2, "text": "P@1", "column_header": true, "row_header": false, "row_section": false}, {"bbox": {"l": 227.16151428222656, "t": 774.0670776367188, "r": 249.1886444091797, "b": 766.1170654296875, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 0, "end_row_offset_idx": 1, "start_col_offset_idx": 2, "end_col_offset_idx": 3, "text": "P@10", "column_header": true, "row_header": false, "row_section": false}, {"bbox": {"l": 258.9349670410156, "t": 774.0670776367188, "r": 285.4084167480469, "b": 766.1170654296875, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 0, "end_row_offset_idx": 1, "start_col_offset_idx": 3, "end_col_offset_idx": 4, "text": "P@100", "column_header": true, "row_header": false, "row_section": false}, {"bbox": {"l": 79.09597778320312, "t": 758.3760375976562, "r": 140.64208984375, "b": 750.426025390625, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 1, "end_row_offset_idx": 2, "start_col_offset_idx": 0, "end_col_offset_idx": 1, "text": "Random retrieval", "column_header": false, "row_header": true, "row_section": false}, {"bbox": {"l": 190.5324249267578, "t": 758.3760375976562, "r": 214.98724365234375, "b": 750.426025390625, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 1, "end_row_offset_idx": 2, "start_col_offset_idx": 1, "end_col_offset_idx": 2, "text": "0.0015", "column_header": false, "row_header": false, "row_section": false}, {"bbox": {"l": 224.7335968017578, "t": 758.3760375976562, "r": 249.18841552734375, "b": 750.426025390625, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 1, "end_row_offset_idx": 2, "start_col_offset_idx": 2, "end_col_offset_idx": 3, "text": "0.0153", "column_header": false, "row_header": false, "row_section": false}, {"bbox": {"l": 260.9533996582031, "t": 758.3760375976562, "r": 285.408203125, "b": 750.426025390625, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 1, "end_row_offset_idx": 2, "start_col_offset_idx": 3, "end_col_offset_idx": 4, "text": "0.1531", "column_header": false, "row_header": false, "row_section": false}, {"bbox": {"l": 79.09597778320312, "t": 747.3314208984375, "r": 180.4634246826172, "b": 739.3814086914062, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 2, "end_row_offset_idx": 3, "start_col_offset_idx": 0, "end_col_offset_idx": 1, "text": "Length-frequency alignment", "column_header": false, "row_header": true, "row_section": false}, {"bbox": {"l": 190.5324249267578, "t": 747.3314819335938, "r": 214.98724365234375, "b": 739.3814697265625, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 2, "end_row_offset_idx": 3, "start_col_offset_idx": 1, "end_col_offset_idx": 2, "text": "0.0032", "column_header": false, "row_header": false, "row_section": false}, {"bbox": {"l": 224.7335968017578, "t": 747.3314819335938, "r": 249.18841552734375, "b": 739.3814697265625, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 2, "end_row_offset_idx": 3, "start_col_offset_idx": 2, "end_col_offset_idx": 3, "text": "0.0127", "column_header": false, "row_header": false, "row_section": false}, {"bbox": {"l": 260.9533996582031, "t": 747.3314819335938, "r": 285.408203125, "b": 739.3814697265625, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 2, "end_row_offset_idx": 3, "start_col_offset_idx": 3, "end_col_offset_idx": 4, "text": "0.6053", "column_header": false, "row_header": false, "row_section": false}, {"bbox": {"l": 79.09597778320312, "t": 736.286865234375, "r": 175.18118286132812, "b": 728.3368530273438, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 3, "end_row_offset_idx": 4, "start_col_offset_idx": 0, "end_col_offset_idx": 1, "text": "Non-isomorphic alignment", "column_header": false, "row_header": true, "row_section": false}, {"bbox": {"l": 190.5324249267578, "t": 736.286865234375, "r": 214.98724365234375, "b": 728.3368530273438, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 3, "end_row_offset_idx": 4, "start_col_offset_idx": 1, "end_col_offset_idx": 2, "text": "0.0000", "column_header": false, "row_header": false, "row_section": false}, {"bbox": {"l": 224.7335968017578, "t": 736.286865234375, "r": 249.18841552734375, "b": 728.3368530273438, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 3, "end_row_offset_idx": 4, "start_col_offset_idx": 2, "end_col_offset_idx": 3, "text": "0.0121", "column_header": false, "row_header": false, "row_section": false}, {"bbox": {"l": 260.9533996582031, "t": 736.286865234375, "r": 285.408203125, "b": 728.3368530273438, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 3, "end_row_offset_idx": 4, "start_col_offset_idx": 3, "end_col_offset_idx": 4, "text": "0.1105", "column_header": false, "row_header": false, "row_section": false}], "num_rows": 4, "num_cols": 4, "grid": [[{"bbox": {"l": 79.09597778320312, "t": 774.0670776367188, "r": 109.7222900390625, "b": 766.1170654296875, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 0, "end_row_offset_idx": 1, "start_col_offset_idx": 0, "end_col_offset_idx": 1, "text": "Baseline", "column_header": true, "row_header": false, "row_section": false}, {"bbox": {"l": 197.4066619873047, "t": 774.0670776367188, "r": 214.98745727539062, "b": 766.1170654296875, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 0, "end_row_offset_idx": 1, "start_col_offset_idx": 1, "end_col_offset_idx": 2, "text": "P@1", "column_header": true, "row_header": false, "row_section": false}, {"bbox": {"l": 227.16151428222656, "t": 774.0670776367188, "r": 249.1886444091797, "b": 766.1170654296875, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 0, "end_row_offset_idx": 1, "start_col_offset_idx": 2, "end_col_offset_idx": 3, "text": "P@10", "column_header": true, "row_header": false, "row_section": false}, {"bbox": {"l": 258.9349670410156, "t": 774.0670776367188, "r": 285.4084167480469, "b": 766.1170654296875, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 0, "end_row_offset_idx": 1, "start_col_offset_idx": 3, "end_col_offset_idx": 4, "text": "P@100", "column_header": true, "row_header": false, "row_section": false}], [{"bbox": {"l": 79.09597778320312, "t": 758.3760375976562, "r": 140.64208984375, "b": 750.426025390625, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 1, "end_row_offset_idx": 2, "start_col_offset_idx": 0, "end_col_offset_idx": 1, "text": "Random retrieval", "column_header": false, "row_header": true, "row_section": false}, {"bbox": {"l": 190.5324249267578, "t": 758.3760375976562, "r": 214.98724365234375, "b": 750.426025390625, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 1, "end_row_offset_idx": 2, "start_col_offset_idx": 1, "end_col_offset_idx": 2, "text": "0.0015", "column_header": false, "row_header": false, "row_section": false}, {"bbox": {"l": 224.7335968017578, "t": 758.3760375976562, "r": 249.18841552734375, "b": 750.426025390625, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 1, "end_row_offset_idx": 2, "start_col_offset_idx": 2, "end_col_offset_idx": 3, "text": "0.0153", "column_header": false, "row_header": false, "row_section": false}, {"bbox": {"l": 260.9533996582031, "t": 758.3760375976562, "r": 285.408203125, "b": 750.426025390625, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 1, "end_row_offset_idx": 2, "start_col_offset_idx": 3, "end_col_offset_idx": 4, "text": "0.1531", "column_header": false, "row_header": false, "row_section": false}], [{"bbox": {"l": 79.09597778320312, "t": 747.3314208984375, "r": 180.4634246826172, "b": 739.3814086914062, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 2, "end_row_offset_idx": 3, "start_col_offset_idx": 0, "end_col_offset_idx": 1, "text": "Length-frequency alignment", "column_header": false, "row_header": true, "row_section": false}, {"bbox": {"l": 190.5324249267578, "t": 747.3314819335938, "r": 214.98724365234375, "b": 739.3814697265625, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 2, "end_row_offset_idx": 3, "start_col_offset_idx": 1, "end_col_offset_idx": 2, "text": "0.0032", "column_header": false, "row_header": false, "row_section": false}, {"bbox": {"l": 224.7335968017578, "t": 747.3314819335938, "r": 249.18841552734375, "b": 739.3814697265625, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 2, "end_row_offset_idx": 3, "start_col_offset_idx": 2, "end_col_offset_idx": 3, "text": "0.0127", "column_header": false, "row_header": false, "row_section": false}, {"bbox": {"l": 260.9533996582031, "t": 747.3314819335938, "r": 285.408203125, "b": 739.3814697265625, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 2, "end_row_offset_idx": 3, "start_col_offset_idx": 3, "end_col_offset_idx": 4, "text": "0.6053", "column_header": false, "row_header": false, "row_section": false}], [{"bbox": {"l": 79.09597778320312, "t": 736.286865234375, "r": 175.18118286132812, "b": 728.3368530273438, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 3, "end_row_offset_idx": 4, "start_col_offset_idx": 0, "end_col_offset_idx": 1, "text": "Non-isomorphic alignment", "column_header": false, "row_header": true, "row_section": false}, {"bbox": {"l": 190.5324249267578, "t": 736.286865234375, "r": 214.98724365234375, "b": 728.3368530273438, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 3, "end_row_offset_idx": 4, "start_col_offset_idx": 1, "end_col_offset_idx": 2, "text": "0.0000", "column_header": false, "row_header": false, "row_section": false}, {"bbox": {"l": 224.7335968017578, "t": 736.286865234375, "r": 249.18841552734375, "b": 728.3368530273438, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 3, "end_row_offset_idx": 4, "start_col_offset_idx": 2, "end_col_offset_idx": 3, "text": "0.0121", "column_header": false, "row_header": false, "row_section": false}, {"bbox": {"l": 260.9533996582031, "t": 736.286865234375, "r": 285.408203125, "b": 728.3368530273438, "coord_origin": "BOTTOMLEFT"}, "row_span": 1, "col_span": 1, "start_row_offset_idx": 3, "end_row_offset_idx": 4, "start_col_offset_idx": 3, "end_col_offset_idx": 4, "text": "0.1105", "column_header": false, "row_header": false, "row_section": false}]]}} | null |
|
9cffc2a5-950c-425e-848f-99d35b459371 | 2302.06555v2.pdf | text | is much more conservative than other metrics used for similar problems, including pairwise matching accuracy, percentile rank, and Pearson correlation (Minnema and Herbelot, 2019). Pairwise matching accuracy and percentile rank have random baseline scores of 0.5, and they converge in the limit. If a has a percentile rank of p in a list A , it will be higher than a random member of A p percent of the time. Pearson correlation is monotonically increasing with pairwise matching accuracy, but P@ k scores are more conservative than any of them for reasonably small values of k . In our case, our target space is 79,059 words, so it is possible to have P@100 values of 0.0 and yet still have near-perfect pairwise matching accuracy, percentile rank, and Pearson correlation scores. P@ k scores also have the advantage that they are intuitive and practically relevant, e.g., for decoding. | null | 441 | 485 | 72 | image/png | 6 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/59", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 5, "bbox": {"l": 71.28717041015625, "t": 664.7708740234375, "r": 292.0834045410156, "b": 422.3381652832031, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 887]}], "orig": "is much more conservative than other metrics used for similar problems, including pairwise matching accuracy, percentile rank, and Pearson correlation (Minnema and Herbelot, 2019). Pairwise matching accuracy and percentile rank have random baseline scores of 0.5, and they converge in the limit. If a has a percentile rank of p in a list A , it will be higher than a random member of A p percent of the time. Pearson correlation is monotonically increasing with pairwise matching accuracy, but P@ k scores are more conservative than any of them for reasonably small values of k . In our case, our target space is 79,059 words, so it is possible to have P@100 values of 0.0 and yet still have near-perfect pairwise matching accuracy, percentile rank, and Pearson correlation scores. P@ k scores also have the advantage that they are intuitive and practically relevant, e.g., for decoding.", "text": "is much more conservative than other metrics used for similar problems, including pairwise matching accuracy, percentile rank, and Pearson correlation (Minnema and Herbelot, 2019). Pairwise matching accuracy and percentile rank have random baseline scores of 0.5, and they converge in the limit. If a has a percentile rank of p in a list A , it will be higher than a random member of A p percent of the time. Pearson correlation is monotonically increasing with pairwise matching accuracy, but P@ k scores are more conservative than any of them for reasonably small values of k . In our case, our target space is 79,059 words, so it is possible to have P@100 values of 0.0 and yet still have near-perfect pairwise matching accuracy, percentile rank, and Pearson correlation scores. P@ k scores also have the advantage that they are intuitive and practically relevant, e.g., for decoding."} | null |
|
5c634071-8d05-4683-9f05-c402eec59de6 | 2302.06555v2.pdf | text | Random retrieval baseline. Our target space of 79,059 words makes the random retrieval baseline: | null | 442 | 49 | 72 | image/png | 6 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/60", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 5, "bbox": {"l": 71.1332015991211, "t": 412.76654052734375, "r": 291.7831726074219, "b": 388.6746520996094, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 96]}], "orig": "Random retrieval baseline. Our target space of 79,059 words makes the random retrieval baseline:", "text": "Random retrieval baseline. Our target space of 79,059 words makes the random retrieval baseline:"} | null |
|
1278ce02-1320-42a5-9cfb-d30565291a10 | 2302.06555v2.pdf | formula | P@ 1 = 1 N N ∑ i =1 n$_{i}$ U (1) | null | 301 | 70 | 72 | image/png | 6 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/61", "parent": {"cref": "#/body"}, "children": [], "label": "formula", "prov": [{"page_no": 5, "bbox": {"l": 140.55914306640625, "t": 377.66351318359375, "r": 290.9989929199219, "b": 342.2218017578125, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 33]}], "orig": "P@ 1 = 1 N N \u2211 i =1 n$_{i}$ U (1)", "text": "P@ 1 = 1 N N \u2211 i =1 n$_{i}$ U (1)"} | null |
|
89af5241-b077-4675-971d-19f36ea358fa | 2302.06555v2.pdf | text | where N represents the total number of image classes; i iterates over each image class; n$_{i}$ denotes the number of labels for image class i ; U refers to the total number of unique aliases. From Equation 1, we get P@1 ≈ 0 . 0015% . | null | 441 | 130 | 72 | image/png | 6 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/62", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 5, "bbox": {"l": 71.25822448730469, "t": 331.0116271972656, "r": 292.0785217285156, "b": 265.8031005859375, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 234]}], "orig": "where N represents the total number of image classes; i iterates over each image class; n$_{i}$ denotes the number of labels for image class i ; U refers to the total number of unique aliases. From Equation 1, we get P@1 \u2248 0 . 0015% .", "text": "where N represents the total number of image classes; i iterates over each image class; n$_{i}$ denotes the number of labels for image class i ; U refers to the total number of unique aliases. From Equation 1, we get P@1 \u2248 0 . 0015% ."} | null |
|
55e1d195-297b-4a19-9e09-b0abb40ec9f9 | 2302.06555v2.pdf | text | Length-frequency alignment baseline. The random retrieval baseline tells us how well we can align representations across the two modalities in the absence of any signal (by chance). However, the fact that we can do better than a random baseline, does not, strictly speaking, prove that our models partially converge toward any sophisticated form of modeling the world. Maybe they simply | null | 442 | 213 | 72 | image/png | 6 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/63", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 5, "bbox": {"l": 71.22360229492188, "t": 255.29852294921875, "r": 292.08270263671875, "b": 149.0504150390625, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 386]}], "orig": "Length-frequency alignment baseline. The random retrieval baseline tells us how well we can align representations across the two modalities in the absence of any signal (by chance). However, the fact that we can do better than a random baseline, does not, strictly speaking, prove that our models partially converge toward any sophisticated form of modeling the world. Maybe they simply", "text": "Length-frequency alignment baseline. The random retrieval baseline tells us how well we can align representations across the two modalities in the absence of any signal (by chance). However, the fact that we can do better than a random baseline, does not, strictly speaking, prove that our models partially converge toward any sophisticated form of modeling the world. Maybe they simply"} | null |
|
315395b8-2614-46b1-9ef5-e562835a112b | 2302.06555v2.pdf | footnote | of a banana into the word ‘banana’, as training pairs to induce a mapping Ω . If Ω then maps the image of a lemon onto the word ‘lemon’ as its nearest neighbor, we say that the precisionat-one for this mapping is 100%. If two target aliases were listed in the bimodal dictionary for the source image, mapping the image onto either of them would result in P@ 1 = 100% . | null | 442 | 130 | 72 | image/png | 6 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/64", "parent": {"cref": "#/body"}, "children": [], "label": "footnote", "prov": [{"page_no": 5, "bbox": {"l": 70.99034881591797, "t": 141.355224609375, "r": 291.7580261230469, "b": 76.57489013671875, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 368]}], "orig": "of a banana into the word \u2018banana\u2019, as training pairs to induce a mapping \u2126 . If \u2126 then maps the image of a lemon onto the word \u2018lemon\u2019 as its nearest neighbor, we say that the precisionat-one for this mapping is 100%. If two target aliases were listed in the bimodal dictionary for the source image, mapping the image onto either of them would result in P@ 1 = 100% .", "text": "of a banana into the word \u2018banana\u2019, as training pairs to induce a mapping \u2126 . If \u2126 then maps the image of a lemon onto the word \u2018lemon\u2019 as its nearest neighbor, we say that the precisionat-one for this mapping is 100%. If two target aliases were listed in the bimodal dictionary for the source image, mapping the image onto either of them would result in P@ 1 = 100% ."} | null |
|
801cd3c0-c84e-4cd1-9371-710f36591cc2 | 2302.06555v2.pdf | caption | Figure 3: t-SNE plot of 5 words mapped from MAE$_{Huge}$ (blue) to OPT$_{30B}$ (orange) using Procrustes analysis. The green represent the mapped MAE$_{Huge}$ embeddings. | null | 442 | 105 | 72 | image/png | 6 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/65", "parent": {"cref": "#/body"}, "children": [], "label": "caption", "prov": [{"page_no": 5, "bbox": {"l": 306.365478515625, "t": 638.8778076171875, "r": 527.3575439453125, "b": 586.4058837890625, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 170]}], "orig": "Figure 3: t-SNE plot of 5 words mapped from MAE$_{Huge}$ (blue) to OPT$_{30B}$ (orange) using Procrustes analysis. The green represent the mapped MAE$_{Huge}$ embeddings.", "text": "Figure 3: t-SNE plot of 5 words mapped from MAE$_{Huge}$ (blue) to OPT$_{30B}$ (orange) using Procrustes analysis. The green represent the mapped MAE$_{Huge}$ embeddings."} | null |
|
00842dd0-5503-45db-93eb-e97de5e70515 | 2302.06555v2.pdf | picture | null | Figure 3: t-SNE plot of 5 words mapped from MAE$_{Huge}$ (blue) to OPT$_{30B}$ (orange) using Procrustes analysis. The green represent the mapped MAE$_{Huge}$ embeddings. | 432 | 255 | 72 | image/png | 6 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/pictures/9", "parent": {"cref": "#/body"}, "children": [], "label": "picture", "prov": [{"page_no": 5, "bbox": {"l": 308.20428466796875, "t": 778.9642944335938, "r": 524.2200927734375, "b": 651.4071655273438, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 170]}], "captions": [{"cref": "#/texts/65"}], "references": [], "footnotes": [], "image": {"mimetype": "image/png", "dpi": 144, "size": {"width": 432.0, "height": 255.0}, "uri": null}, "annotations": []} | null |
|
a114e3d5-0d2c-4bc9-96a2-f3c4388fe4b6 | 2302.06555v2.pdf | text | pick up on shallow characteristics shared across the two spaces. One example is frequency: frequent words may refer to frequently depicted objects. Learning what is rare is learning about the world, but more is at stake in the debate around whether LMs understand. Or consider length: word length may correlate with the structural complexity of objects (in some way), and maybe this is what drives our alignment precision? To control for such effects, we run a second baseline aligning representations from computer vision models to two-dimensional word representations, representing words by their length and frequency. We collected frequency data based on English Wikipedia using NLTK (Bird et al., 2009) for all aliases within our target space. We use PCA and Procrustes Analysis or ridge regression (Toneva and Wehbe, 2019) to map into the length-frequency space and report the best of those as a second, stronger baseline. | null | 443 | 509 | 72 | image/png | 6 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/66", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 5, "bbox": {"l": 306.16961669921875, "t": 560.0623779296875, "r": 527.3591918945312, "b": 305.34613037109375, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 927]}], "orig": "pick up on shallow characteristics shared across the two spaces. One example is frequency: frequent words may refer to frequently depicted objects. Learning what is rare is learning about the world, but more is at stake in the debate around whether LMs understand. Or consider length: word length may correlate with the structural complexity of objects (in some way), and maybe this is what drives our alignment precision? To control for such effects, we run a second baseline aligning representations from computer vision models to two-dimensional word representations, representing words by their length and frequency. We collected frequency data based on English Wikipedia using NLTK (Bird et al., 2009) for all aliases within our target space. We use PCA and Procrustes Analysis or ridge regression (Toneva and Wehbe, 2019) to map into the length-frequency space and report the best of those as a second, stronger baseline.", "text": "pick up on shallow characteristics shared across the two spaces. One example is frequency: frequent words may refer to frequently depicted objects. Learning what is rare is learning about the world, but more is at stake in the debate around whether LMs understand. Or consider length: word length may correlate with the structural complexity of objects (in some way), and maybe this is what drives our alignment precision? To control for such effects, we run a second baseline aligning representations from computer vision models to two-dimensional word representations, representing words by their length and frequency. We collected frequency data based on English Wikipedia using NLTK (Bird et al., 2009) for all aliases within our target space. We use PCA and Procrustes Analysis or ridge regression (Toneva and Wehbe, 2019) to map into the length-frequency space and report the best of those as a second, stronger baseline."} | null |
|
e58f7f49-4941-4c85-a409-6e922bdc74fb | 2302.06555v2.pdf | text | Non-isomorphic alignment baseline. The former two baselines examine the possibility of aligning representations across two modalities based on chance or shallow signals. While informative, neither strictly demonstrates that a linear projection cannot effectively establish a connection between two non-isomorphic representation spaces, potentially outperforming the random or lengthfrequency baselines. To rigorously explore this, we disrupt the relationship between words and their corresponding representations by shuffling them. This permutation ensures that the source and target spaces become non-isomorphic. Specifically, we shuffled OPT$_{30B}$ three times at random and report the alignment results between those and original OPT$_{30B}$, we use the same Procrustes analysis for | null | 443 | 428 | 72 | image/png | 6 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/67", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 5, "bbox": {"l": 306.20733642578125, "t": 289.99688720703125, "r": 527.4515380859375, "b": 76.0753173828125, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 786]}], "orig": "Non-isomorphic alignment baseline. The former two baselines examine the possibility of aligning representations across two modalities based on chance or shallow signals. While informative, neither strictly demonstrates that a linear projection cannot effectively establish a connection between two non-isomorphic representation spaces, potentially outperforming the random or lengthfrequency baselines. To rigorously explore this, we disrupt the relationship between words and their corresponding representations by shuffling them. This permutation ensures that the source and target spaces become non-isomorphic. Specifically, we shuffled OPT$_{30B}$ three times at random and report the alignment results between those and original OPT$_{30B}$, we use the same Procrustes analysis for", "text": "Non-isomorphic alignment baseline. The former two baselines examine the possibility of aligning representations across two modalities based on chance or shallow signals. While informative, neither strictly demonstrates that a linear projection cannot effectively establish a connection between two non-isomorphic representation spaces, potentially outperforming the random or lengthfrequency baselines. To rigorously explore this, we disrupt the relationship between words and their corresponding representations by shuffling them. This permutation ensures that the source and target spaces become non-isomorphic. Specifically, we shuffled OPT$_{30B}$ three times at random and report the alignment results between those and original OPT$_{30B}$, we use the same Procrustes analysis for"} | null |
|
46f07d41-0be5-4d86-9cf0-7991da5d9684 | 2302.06555v2.pdf | caption | Figure 4: LMs converge toward the geometry of visual models as they grow larger on Exclude-1K set. | null | 893 | 21 | 72 | image/png | 7 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/68", "parent": {"cref": "#/body"}, "children": [], "label": "caption", "prov": [{"page_no": 6, "bbox": {"l": 73.5337905883789, "t": 389.9848327636719, "r": 519.9508056640625, "b": 379.31463623046875, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 98]}], "orig": "Figure 4: LMs converge toward the geometry of visual models as they grow larger on Exclude-1K set.", "text": "Figure 4: LMs converge toward the geometry of visual models as they grow larger on Exclude-1K set."} | null |
|
0ab8b3df-d493-4755-90ce-a662ff499418 | 2302.06555v2.pdf | picture | null | Figure 4: LMs converge toward the geometry of visual models as they grow larger on Exclude-1K set. | 853 | 689 | 72 | image/png | 7 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/pictures/10", "parent": {"cref": "#/body"}, "children": [], "label": "picture", "prov": [{"page_no": 6, "bbox": {"l": 77.1706314086914, "t": 752.3418579101562, "r": 503.3846435546875, "b": 408.00994873046875, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 98]}], "captions": [{"cref": "#/texts/68"}], "references": [], "footnotes": [], "image": {"mimetype": "image/png", "dpi": 144, "size": {"width": 853.0, "height": 689.0}, "uri": null}, "annotations": []} | null |
|
1a77051c-a165-494b-ba02-f228783e0409 | 2302.06555v2.pdf | text | computing the alignment. Table 2 presents a comparison of the three different baselines. All baselines have P@100 well below 1%. Our mappings between VMs and LMs score much higher (up to 64%), showing the strength of the correlation between the geometries induced by these models with respect to a conservative performance metric. | null | 441 | 185 | 72 | image/png | 7 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/69", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 6, "bbox": {"l": 71.34822082519531, "t": 354.8316955566406, "r": 292.0813903808594, "b": 262.27301025390625, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 330]}], "orig": "computing the alignment. Table 2 presents a comparison of the three different baselines. All baselines have P@100 well below 1%. Our mappings between VMs and LMs score much higher (up to 64%), showing the strength of the correlation between the geometries induced by these models with respect to a conservative performance metric.", "text": "computing the alignment. Table 2 presents a comparison of the three different baselines. All baselines have P@100 well below 1%. Our mappings between VMs and LMs score much higher (up to 64%), showing the strength of the correlation between the geometries induced by these models with respect to a conservative performance metric."} | null |
|
84a88245-492a-4513-8107-e1c0ba51c59d | 2302.06555v2.pdf | section_header | 5 Results | null | 112 | 23 | 72 | image/png | 7 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/70", "parent": {"cref": "#/body"}, "children": [], "label": "section_header", "prov": [{"page_no": 6, "bbox": {"l": 71.40553283691406, "t": 246.7689208984375, "r": 127.26032257080078, "b": 235.4713592529297, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 9]}], "orig": "5 Results", "text": "5 Results", "level": 1} | null |
|
8442b477-b8e0-451c-83fe-54f5c4bac58c | 2302.06555v2.pdf | text | Similarities between visual and textual representations and how they are recovered through Procrustes Analysis are visualized through t-SNE in Figure 3. Our main results for nine VMs and all LMs are presented in Figure 4. The best P@100 scores are around 64%, with baseline scores lower than 1% (Table 2). In general, even the smallest language models outperform the baselines by orders of magnitude. We focus mainly on P@10 and P@100 scores because P@1 only allows one surface form to express a visual concept, but in reality, | null | 442 | 295 | 72 | image/png | 7 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/71", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 6, "bbox": {"l": 71.2156753540039, "t": 222.7291259765625, "r": 292.0754089355469, "b": 75.686279296875, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 527]}], "orig": "Similarities between visual and textual representations and how they are recovered through Procrustes Analysis are visualized through t-SNE in Figure 3. Our main results for nine VMs and all LMs are presented in Figure 4. The best P@100 scores are around 64%, with baseline scores lower than 1% (Table 2). In general, even the smallest language models outperform the baselines by orders of magnitude. We focus mainly on P@10 and P@100 scores because P@1 only allows one surface form to express a visual concept, but in reality,", "text": "Similarities between visual and textual representations and how they are recovered through Procrustes Analysis are visualized through t-SNE in Figure 3. Our main results for nine VMs and all LMs are presented in Figure 4. The best P@100 scores are around 64%, with baseline scores lower than 1% (Table 2). In general, even the smallest language models outperform the baselines by orders of magnitude. We focus mainly on P@10 and P@100 scores because P@1 only allows one surface form to express a visual concept, but in reality,"} | null |
|
e4b5caf4-460f-4b2f-af09-49e178d47915 | 2302.06555v2.pdf | text | an artifact such as a vehicle may be denoted by many lexemes (car, automobile, SUV, etc.), each of which may have multiple inflections and derivations (car, cars, car's, etc.). Figure 5 shows examples where the top predictions seem 'as good' as the gold standard. We find that a region of 10 neighbours corresponds roughly to grammatical forms or synonyms, and a neighbourhood of 100 word forms corresponds roughly to coarse-grained semantic classes. Results of P@10 in Figure 4, show that up to one in five of all visual concepts were mapped to the correct region of the language space, with only a slight deviation from the specific surface form. Considering P@100, we see that more than two thirds of the visual concepts find a semantic match in the language space when using ResNet152 and OPT or LLaMA-2, for example. We see that ResNet models score highest overall, followed by SegFormers, while MAE models rank third. We presume that this ranking is the result, in | null | 442 | 536 | 72 | image/png | 7 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/72", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 6, "bbox": {"l": 306.32073974609375, "t": 354.6774597167969, "r": 527.4515380859375, "b": 86.65802001953125, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 970]}], "orig": "an artifact such as a vehicle may be denoted by many lexemes (car, automobile, SUV, etc.), each of which may have multiple inflections and derivations (car, cars, car's, etc.). Figure 5 shows examples where the top predictions seem 'as good' as the gold standard. We find that a region of 10 neighbours corresponds roughly to grammatical forms or synonyms, and a neighbourhood of 100 word forms corresponds roughly to coarse-grained semantic classes. Results of P@10 in Figure 4, show that up to one in five of all visual concepts were mapped to the correct region of the language space, with only a slight deviation from the specific surface form. Considering P@100, we see that more than two thirds of the visual concepts find a semantic match in the language space when using ResNet152 and OPT or LLaMA-2, for example. We see that ResNet models score highest overall, followed by SegFormers, while MAE models rank third. We presume that this ranking is the result, in", "text": "an artifact such as a vehicle may be denoted by many lexemes (car, automobile, SUV, etc.), each of which may have multiple inflections and derivations (car, cars, car's, etc.). Figure 5 shows examples where the top predictions seem 'as good' as the gold standard. We find that a region of 10 neighbours corresponds roughly to grammatical forms or synonyms, and a neighbourhood of 100 word forms corresponds roughly to coarse-grained semantic classes. Results of P@10 in Figure 4, show that up to one in five of all visual concepts were mapped to the correct region of the language space, with only a slight deviation from the specific surface form. Considering P@100, we see that more than two thirds of the visual concepts find a semantic match in the language space when using ResNet152 and OPT or LLaMA-2, for example. We see that ResNet models score highest overall, followed by SegFormers, while MAE models rank third. We presume that this ranking is the result, in"} | null |
|
2d965b10-92ed-40d9-bbed-0865ffec6183 | 2302.06555v2.pdf | paragraph | Image Classes | null | 118 | 20 | 72 | image/png | 8 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/73", "parent": {"cref": "#/body"}, "children": [], "label": "paragraph", "prov": [{"page_no": 7, "bbox": {"l": 71.53941345214844, "t": 774.43701171875, "r": 130.51449584960938, "b": 764.5503540039062, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 13]}], "orig": "Image Classes", "text": "Image Classes"} | null |
|
50c5d387-e5f5-4af4-b690-5fd28eb9a8ae | 2302.06555v2.pdf | section_header | Nearest Neighbors (Top 100) | null | 233 | 20 | 72 | image/png | 8 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/74", "parent": {"cref": "#/body"}, "children": [], "label": "section_header", "prov": [{"page_no": 7, "bbox": {"l": 195.791259765625, "t": 774.43701171875, "r": 312.44464111328125, "b": 764.481201171875, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 27]}], "orig": "Nearest Neighbors (Top 100)", "text": "Nearest Neighbors (Top 100)", "level": 1} | null |
|
a44f83d5-3acc-4bee-a5c9-721fedfa3748 | 2302.06555v2.pdf | picture | null | null | 119 | 101 | 72 | image/png | 8 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/pictures/11", "parent": {"cref": "#/body"}, "children": [], "label": "picture", "prov": [{"page_no": 7, "bbox": {"l": 71.45142364501953, "t": 762.3618774414062, "r": 131.0451202392578, "b": 712.0277099609375, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 0]}], "captions": [], "references": [], "footnotes": [], "image": {"mimetype": "image/png", "dpi": 144, "size": {"width": 119.0, "height": 101.0}, "uri": null}, "annotations": []} | null |
|
aa678786-5264-43ad-a1ef-2f3a7e9f2b7e | 2302.06555v2.pdf | picture | null | null | 119 | 109 | 72 | image/png | 8 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/pictures/12", "parent": {"cref": "#/body"}, "children": [], "label": "picture", "prov": [{"page_no": 7, "bbox": {"l": 71.35606384277344, "t": 704.283935546875, "r": 130.9281768798828, "b": 650.0250244140625, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 0]}], "captions": [], "references": [], "footnotes": [], "image": {"mimetype": "image/png", "dpi": 144, "size": {"width": 119.0, "height": 109.0}, "uri": null}, "annotations": []} | null |
|
f17e02d8-34f7-457b-96c8-91c8487bc4d5 | 2302.06555v2.pdf | picture | null | null | 119 | 101 | 72 | image/png | 8 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/pictures/13", "parent": {"cref": "#/body"}, "children": [], "label": "picture", "prov": [{"page_no": 7, "bbox": {"l": 71.25289154052734, "t": 645.5927734375, "r": 131.10714721679688, "b": 594.8673095703125, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 0]}], "captions": [], "references": [], "footnotes": [], "image": {"mimetype": "image/png", "dpi": 144, "size": {"width": 119.0, "height": 101.0}, "uri": null}, "annotations": []} | null |
|
2a4dbed3-c4be-46e1-918b-042f4b54da0d | 2302.06555v2.pdf | picture | null | null | 119 | 100 | 72 | image/png | 8 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/pictures/14", "parent": {"cref": "#/body"}, "children": [], "label": "picture", "prov": [{"page_no": 7, "bbox": {"l": 71.36942291259766, "t": 587.2476806640625, "r": 130.98825073242188, "b": 537.20166015625, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 0]}], "captions": [], "references": [], "footnotes": [], "image": {"mimetype": "image/png", "dpi": 144, "size": {"width": 119.0, "height": 100.0}, "uri": null}, "annotations": []} | null |
|
985b2396-eabc-4d66-9946-e641632253c2 | 2302.06555v2.pdf | text | palmyra, palmyra palm, palm, palais, palatines, royal palm , palazzi, palazzo, palisades, palatinate, regency, palatial, palas, palatinates, palms, palimony, caribe, palmier, paladins, banyan tree, bermudas, bruneian, palazzos, bahamian, palmers, malacca, madeira, ceiba tree, palmettos, palmtop, oil palm, pal, royal, regal, roystonea regia , lindens, palaces, athenaeum, arboricultural, gabonese, palming, sugar palm, elm tree, palings, palm tree, palaeography, coconut palm, palisaded, bahraini, nicaraguan, … … , regent, myrtle, estancia, pavonia, imperial, royalist, regnal, historic, annals, maduro, rozelle, dominical, hydropathic, andorran | null | 639 | 95 | 72 | image/png | 8 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/75", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 7, "bbox": {"l": 195.43907165527344, "t": 760.7074584960938, "r": 514.689697265625, "b": 713.5289916992188, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 647]}], "orig": "palmyra, palmyra palm, palm, palais, palatines, royal palm , palazzi, palazzo, palisades, palatinate, regency, palatial, palas, palatinates, palms, palimony, caribe, palmier, paladins, banyan tree, bermudas, bruneian, palazzos, bahamian, palmers, malacca, madeira, ceiba tree, palmettos, palmtop, oil palm, pal, royal, regal, roystonea regia , lindens, palaces, athenaeum, arboricultural, gabonese, palming, sugar palm, elm tree, palings, palm tree, palaeography, coconut palm, palisaded, bahraini, nicaraguan, \u2026 \u2026 , regent, myrtle, estancia, pavonia, imperial, royalist, regnal, historic, annals, maduro, rozelle, dominical, hydropathic, andorran", "text": "palmyra, palmyra palm, palm, palais, palatines, royal palm , palazzi, palazzo, palisades, palatinate, regency, palatial, palas, palatinates, palms, palimony, caribe, palmier, paladins, banyan tree, bermudas, bruneian, palazzos, bahamian, palmers, malacca, madeira, ceiba tree, palmettos, palmtop, oil palm, pal, royal, regal, roystonea regia , lindens, palaces, athenaeum, arboricultural, gabonese, palming, sugar palm, elm tree, palings, palm tree, palaeography, coconut palm, palisaded, bahraini, nicaraguan, \u2026 \u2026 , regent, myrtle, estancia, pavonia, imperial, royalist, regnal, historic, annals, maduro, rozelle, dominical, hydropathic, andorran"} | null |
|
402ad2ae-7330-49bf-b722-cf3f2833f75b | 2302.06555v2.pdf | text | drinking fountain , water fountain , cesspools, water cooler, manhole cover, bird feeder, birdbath, water jug, drainage system, fountain, water tap, watering can, garbage disposal, cesspit, recycling bin, water tank, garbage can, water pipe, manhole, toilet bowl, water closet, cement mixer, trash bin, soda fountain, bubblers, ice chest, footstone, ice machine, churns, milk float, overflowing, privies, grate, disposal, bathing, water bed, trickles, waterworks, drinking vessel, wading pool, carafe, vending machine, toilet water, sandboxes, toilet seat, drainpipe, draining, … … , spring water, ice maker, retaining wall, charcoal burner, litter, sentry box, cistern, waterhole, manholes, baptismal font, waterless | null | 636 | 95 | 72 | image/png | 8 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/76", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 7, "bbox": {"l": 195.55520629882812, "t": 702.2905883789062, "r": 513.342041015625, "b": 654.7896728515625, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 717]}], "orig": "drinking fountain , water fountain , cesspools, water cooler, manhole cover, bird feeder, birdbath, water jug, drainage system, fountain, water tap, watering can, garbage disposal, cesspit, recycling bin, water tank, garbage can, water pipe, manhole, toilet bowl, water closet, cement mixer, trash bin, soda fountain, bubblers, ice chest, footstone, ice machine, churns, milk float, overflowing, privies, grate, disposal, bathing, water bed, trickles, waterworks, drinking vessel, wading pool, carafe, vending machine, toilet water, sandboxes, toilet seat, drainpipe, draining, \u2026 \u2026 , spring water, ice maker, retaining wall, charcoal burner, litter, sentry box, cistern, waterhole, manholes, baptismal font, waterless", "text": "drinking fountain , water fountain , cesspools, water cooler, manhole cover, bird feeder, birdbath, water jug, drainage system, fountain, water tap, watering can, garbage disposal, cesspit, recycling bin, water tank, garbage can, water pipe, manhole, toilet bowl, water closet, cement mixer, trash bin, soda fountain, bubblers, ice chest, footstone, ice machine, churns, milk float, overflowing, privies, grate, disposal, bathing, water bed, trickles, waterworks, drinking vessel, wading pool, carafe, vending machine, toilet water, sandboxes, toilet seat, drainpipe, draining, \u2026 \u2026 , spring water, ice maker, retaining wall, charcoal burner, litter, sentry box, cistern, waterhole, manholes, baptismal font, waterless"} | null |
|
5635b7b6-31c2-46b3-bd21-153192905bcc | 2302.06555v2.pdf | text | clamp, wrench, screwdriver, socket wrench, carabiner , torque wrench, screwdrivers, fastener, elastic bandage, pliers, retractor, screw thread, carabiners, plunger, spanner, corer, screw, aspirator, clamps, adjustable spanner, applicator, center punch, latch, extractor, lever, adaptor, hose, gripper, compensator, pipe wrench, power drill, retractors, bicycle pump, holding device, grappling hook, fasteners, extension cord, locknuts, bungee cord, drill press, ratcheting, elastic band, reamer, soldering iron, handlebar, plug, stopper knot, tongs, twist drill, crimpers, … … , shock absorber, caliper, shackle, wristband, reducer, wrenches, loop knot, safety belt | null | 634 | 95 | 72 | image/png | 8 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/77", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 7, "bbox": {"l": 195.4573974609375, "t": 643.7977905273438, "r": 512.2069091796875, "b": 596.356689453125, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 665]}], "orig": "clamp, wrench, screwdriver, socket wrench, carabiner , torque wrench, screwdrivers, fastener, elastic bandage, pliers, retractor, screw thread, carabiners, plunger, spanner, corer, screw, aspirator, clamps, adjustable spanner, applicator, center punch, latch, extractor, lever, adaptor, hose, gripper, compensator, pipe wrench, power drill, retractors, bicycle pump, holding device, grappling hook, fasteners, extension cord, locknuts, bungee cord, drill press, ratcheting, elastic band, reamer, soldering iron, handlebar, plug, stopper knot, tongs, twist drill, crimpers, \u2026 \u2026 , shock absorber, caliper, shackle, wristband, reducer, wrenches, loop knot, safety belt", "text": "clamp, wrench, screwdriver, socket wrench, carabiner , torque wrench, screwdrivers, fastener, elastic bandage, pliers, retractor, screw thread, carabiners, plunger, spanner, corer, screw, aspirator, clamps, adjustable spanner, applicator, center punch, latch, extractor, lever, adaptor, hose, gripper, compensator, pipe wrench, power drill, retractors, bicycle pump, holding device, grappling hook, fasteners, extension cord, locknuts, bungee cord, drill press, ratcheting, elastic band, reamer, soldering iron, handlebar, plug, stopper knot, tongs, twist drill, crimpers, \u2026 \u2026 , shock absorber, caliper, shackle, wristband, reducer, wrenches, loop knot, safety belt"} | null |
|
02fdcde5-29ec-4075-beb5-e487b6f1d1af | 2302.06555v2.pdf | text | community center, training school, school, youth hostel, service department, conference center, music school, day school, student union , academy, life office, hall, orphanage, school system, meeting, college, ministry, school principal, government building, house, council, clinic, business office, schoolmaster, workshop, council board, boardinghouse, club, service club, schools, detention centre, gymnasium, gym, schoolmasters, … … , nursing home, meeting house, church, education, reform school, semester, schoolmate, study hall, member, schoolrooms, assembly hall, meetings, hotel, district manager, arena, staff member, firm | null | 643 | 94 | 72 | image/png | 8 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/78", "parent": {"cref": "#/body"}, "children": [], "label": "text", "prov": [{"page_no": 7, "bbox": {"l": 195.45948791503906, "t": 585.3809204101562, "r": 516.7018432617188, "b": 538.2024536132812, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 631]}], "orig": "community center, training school, school, youth hostel, service department, conference center, music school, day school, student union , academy, life office, hall, orphanage, school system, meeting, college, ministry, school principal, government building, house, council, clinic, business office, schoolmaster, workshop, council board, boardinghouse, club, service club, schools, detention centre, gymnasium, gym, schoolmasters, \u2026 \u2026 , nursing home, meeting house, church, education, reform school, semester, schoolmate, study hall, member, schoolrooms, assembly hall, meetings, hotel, district manager, arena, staff member, firm", "text": "community center, training school, school, youth hostel, service department, conference center, music school, day school, student union , academy, life office, hall, orphanage, school system, meeting, college, ministry, school principal, government building, house, council, clinic, business office, schoolmaster, workshop, council board, boardinghouse, club, service club, schools, detention centre, gymnasium, gym, schoolmasters, \u2026 \u2026 , nursing home, meeting house, church, education, reform school, semester, schoolmate, study hall, member, schoolrooms, assembly hall, meetings, hotel, district manager, arena, staff member, firm"} | null |
|
df877eec-5049-4528-b781-622bae4f52bb | 2302.06555v2.pdf | picture | null | null | 120 | 101 | 72 | image/png | 8 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/pictures/15", "parent": {"cref": "#/body"}, "children": [], "label": "picture", "prov": [{"page_no": 7, "bbox": {"l": 132.97372436523438, "t": 762.504150390625, "r": 192.79490661621094, "b": 712.1260375976562, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 0]}], "captions": [], "references": [], "footnotes": [], "image": {"mimetype": "image/png", "dpi": 144, "size": {"width": 120.0, "height": 101.0}, "uri": null}, "annotations": []} | null |
|
c4b4432c-40b2-40db-8e20-c6d1c3c4c07e | 2302.06555v2.pdf | picture | null | null | 120 | 109 | 72 | image/png | 8 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/pictures/16", "parent": {"cref": "#/body"}, "children": [], "label": "picture", "prov": [{"page_no": 7, "bbox": {"l": 132.55625915527344, "t": 704.5750732421875, "r": 192.52426147460938, "b": 649.9854125976562, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 0]}], "captions": [], "references": [], "footnotes": [], "image": {"mimetype": "image/png", "dpi": 144, "size": {"width": 120.0, "height": 109.0}, "uri": null}, "annotations": []} | null |
|
4feea742-44e4-42a3-9b88-999e8919e6ef | 2302.06555v2.pdf | picture | null | null | 119 | 101 | 72 | image/png | 8 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/pictures/17", "parent": {"cref": "#/body"}, "children": [], "label": "picture", "prov": [{"page_no": 7, "bbox": {"l": 132.8827362060547, "t": 645.2271118164062, "r": 192.2962646484375, "b": 594.8204956054688, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 0]}], "captions": [], "references": [], "footnotes": [], "image": {"mimetype": "image/png", "dpi": 144, "size": {"width": 119.0, "height": 101.0}, "uri": null}, "annotations": []} | null |
|
80a9e68e-cf48-473c-9b83-3729fc5b71ac | 2302.06555v2.pdf | caption | Figure 5: Examples featuring the 100 nearest neighbors in the mapping of image classes into the language representation space (from MAE$_{Huge}$ to OPT$_{30B}$). The golden labels are highlighted in green. | null | 908 | 49 | 72 | image/png | 8 | application/pdf | 1.0.0 | [] | null | null | {"self_ref": "#/texts/79", "parent": {"cref": "#/body"}, "children": [], "label": "caption", "prov": [{"page_no": 7, "bbox": {"l": 71.26412200927734, "t": 521.2467041015625, "r": 525.5408325195312, "b": 497.1179504394531, "coord_origin": "BOTTOMLEFT"}, "charspan": [0, 205]}], "orig": "Figure 5: Examples featuring the 100 nearest neighbors in the mapping of image classes into the language representation space (from MAE$_{Huge}$ to OPT$_{30B}$). The golden labels are highlighted in green.", "text": "Figure 5: Examples featuring the 100 nearest neighbors in the mapping of image classes into the language representation space (from MAE$_{Huge}$ to OPT$_{30B}$). The golden labels are highlighted in green."} | null |