File size: 36,929 Bytes
e4342db |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 476 477 478 479 480 481 482 483 484 485 486 487 488 489 490 491 492 493 494 495 496 497 498 499 500 501 502 503 504 505 506 507 508 509 510 511 512 513 514 515 516 517 518 519 520 521 522 523 524 525 526 527 528 529 530 531 532 533 534 535 536 537 538 539 540 541 542 543 544 545 546 547 548 549 550 551 552 553 554 555 556 557 558 559 560 561 562 563 564 565 566 567 568 569 570 571 572 573 574 575 576 577 578 579 580 581 582 583 584 585 586 587 588 589 590 591 592 593 594 595 596 597 598 599 600 601 602 603 604 605 606 607 608 609 610 611 612 613 614 615 616 617 618 619 620 621 622 623 624 625 626 627 628 629 630 631 632 633 634 635 636 637 638 639 640 641 642 643 644 645 646 647 648 649 650 651 652 653 654 655 656 657 658 659 660 661 662 663 664 665 666 667 668 669 670 671 672 673 674 675 676 677 678 679 680 681 682 683 684 685 686 687 688 689 690 691 692 693 694 695 696 697 698 699 700 701 702 703 704 705 706 707 708 709 710 711 712 713 714 715 716 717 718 719 720 721 722 723 724 725 726 727 728 729 730 731 732 733 734 735 736 737 738 739 740 741 742 743 744 745 746 747 748 749 750 751 752 753 754 755 756 757 758 759 760 761 762 763 764 765 766 767 768 769 770 771 772 773 774 775 776 777 778 779 780 781 782 783 784 785 786 787 788 789 790 791 792 793 794 795 796 797 798 799 800 801 802 803 804 805 806 807 808 809 810 811 812 813 814 815 816 817 818 819 820 821 822 823 824 825 826 827 828 829 830 831 832 833 834 835 836 837 838 839 840 841 842 843 844 845 846 847 848 849 850 851 852 853 854 855 856 857 858 859 860 861 862 863 864 865 866 867 868 869 870 871 872 873 874 875 876 877 878 879 880 881 882 883 884 885 886 887 888 889 890 891 892 893 894 895 896 897 898 899 900 901 902 903 904 905 906 907 908 909 910 911 912 913 914 915 916 917 918 919 920 921 922 923 924 925 926 927 928 929 930 931 932 933 934 935 936 937 938 939 940 941 942 943 944 945 946 947 948 949 950 951 952 953 954 955 956 957 958 959 960 961 962 963 964 965 966 967 968 969 970 971 972 973 974 975 976 977 978 979 980 981 982 983 984 985 986 987 988 989 990 991 992 993 994 995 996 997 998 999 1000 1001 1002 1003 1004 1005 1006 1007 1008 1009 1010 1011 1012 1013 1014 1015 1016 1017 1018 1019 1020 1021 1022 1023 1024 1025 1026 1027 1028 1029 1030 1031 1032 1033 1034 1035 1036 1037 1038 1039 1040 1041 1042 1043 1044 1045 1046 1047 1048 1049 1050 1051 1052 1053 1054 1055 1056 1057 1058 1059 1060 1061 1062 1063 1064 1065 1066 1067 1068 1069 1070 1071 1072 1073 1074 1075 1076 1077 1078 1079 1080 1081 1082 1083 1084 1085 1086 1087 1088 1089 1090 1091 1092 1093 1094 1095 1096 1097 1098 1099 1100 1101 |
Image Memorability Prediction with Vision Transformers Thomas Hagen1,� and Thomas Espeseth1,2 1Department of Psychology, University of Oslo, Oslo, Norway 2Department of Psychology, Oslo New University College, Oslo, Norway Behavioral studies have shown that the memorability of images is similar across groups of people, suggesting that memorability is a function of the intrinsic properties of images, and is unre- lated to people’s individual experiences and traits. Deep learn- ing networks can be trained on such properties and be used to predict memorability in new data sets. Convolutional neu- ral networks (CNN) have pioneered image memorability predic- tion, but more recently developed vision transformer (ViT) mod- els may have the potential to yield even better predictions. In this paper, we present the ViTMem, a new memorability model based on ViT, and evaluate memorability predictions obtained by it with state-of-the-art CNN-derived models. Results showed that ViTMem performed equal to or better than state-of-the- art models on all data sets. Additional semantic level analyses revealed that ViTMem is particularly sensitive to the seman- tic content that drives memorability in images. We conclude that ViTMem provides a new step forward, and propose that ViT-derived models can replace CNNs for computational pre- diction of image memorability. Researchers, educators, adver- tisers, visual designers and other interested parties can leverage the model to improve the memorability of their image material. memorability | vision transformers | psychology | semantic information Introduction Everyone knows that our memories depend on the experi- ences we have had, facts we have encountered, and the abil- ities we have to remember them. Combinations of these fac- tors differ between individuals and give rise to unique memo- ries in each of us. However, a complementary perspective on memory focuses on the material that is (to be) remembered rather than the individual that does the remembering. In one central study, Isola et al. (1) presented more than 2000 scene images in a continuous repeat-detection task. The partici- pants were asked to respond whenever they saw an identical repeat. The results revealed that the memorability score (per- cent correct detections) varied considerably between images. Most importantly, by running a consistency analysis in which Spearman’s rank correlation was calculated on the memo- rability scores from random splits of the participant group, Isola and colleagues (1) were able to show that the memora- bility score ranking was consistent across participants – some images were memorable and some were forgettable. These results indicate that the degree to which an image was cor- rectly detected depended on properties intrinsic to the image itself, not the traits of the observers. This is important be- cause it shows that one can use the memorability scores in a stimulus set to predict memory performance in a new group of participants. These results have been replicated and extended in a num- ber of studies, revealing that similar findings are obtained with different memory tasks (2), different retention times (1, 2), different contexts (3), and independent of whether en- coding is intentional or incidental (4). However, although image memorability has proven to be a robust and reliable phenomenon, it has not been straightforward to pinpoint the image properties that drive it. What seems clear though, is that memorability is multifaceted (5, 6). One way to char- acterize the underpinnings of memorability is to investigate the contribution from processes at different levels of the vi- sual processing stream. For example, at the earliest stages of processing of a visual scene, visual attributes such as local contrast, orientation, and color are coded. At an intermedi- ate level, contours are integrated, surfaces, shapes, and depth cues are segmented, and foreground and background are dis- tinguished. At a higher level, object recognition is conducted through matching with templates stored in long term mem- ory. Positive correlations between brightness and high contrast of objects with memorability has been found (7), but in general, low-level visual factors such as color, contrast, and spatial frequency do not predict memorability well (5, 8, 9). This is consistent with results showing that perceptual features are typically not retained in long term visual memory (10). In contrast to the low-level features, the evidence for a re- lation between intermediate to high level semantic features and memorability is much stronger. For example, images that contain people, faces, body parts, animals, and food are often associated with high memorability, whereas the opposite is a typical finding for objects like buildings and furniture and images of landscapes and parks (3, 7, 11, 12). Other inter- mediate to high level features such as object interaction with the context or other objects, saliency factors, and image com- position also contribute to memorability (5). Furthermore, although memorability is not reducible to high-level features such as aesthetics (1, 12), interestingness (1, 13), or popu- larity (12), emotions, particularly of negative valence, seem to predict higher memorability (9, 12). Finally, memorabil- ity seems to be relatively independent of cognitive control, attention, or priming (14). Overall, the available evidence indicates that memorability seems to capture intermediate- to high-level properties of semantics, such as objects or actions, and image composi- tion, such as layout and clutter, rather than low-level fea- Hagen et al. | January 23, 2023 | 1–7 arXiv:2301.08647v1 [cs.CV] 20 Jan 2023 tures (5, 15). This fits well with the central role of semantic categories in organizing cognition and memory (16). Gen- erally, the priority of semantic-level information enables us to quickly understand novel scenes and predict future events (17). For example, when inspecting a novel scene or an im- age, we do not primarily focus on low-level perceptual fea- tures or pixels, but prioritize more abstract visual schemas involving spatial regions, objects, and the relation between them (18). Also, when people are asked to indicate which regions of an image helps them recognize an image, there is high consistency between people’s responses (18). Similarly, fixation map data from eye-tracking have shown that there is a positive correlation between fixation map consistency and scene memorability, and this relation is associated with the presence of meaningful objects (3, 7, 19). Bylinskii et al. (5) suggest that these properties most efficiently signal infor- mation of high utility to our species, for example, emotions, social aspects, animate objects (e.g., faces, gestures, interac- tions), unexpected events, and tangible objects. Memorability prediction. The finding that the memorabil- ity of an image is governed by properties intrinsic to the im- age itself, not only implies that one can predict memory per- formance in a new set of participants, as described above, but also that one can predict the memorability of a novel set of images (i.e., memorability is an “image computable” fea- ture). Given the availability of computational algorithms and high-quality training sets of sufficient size, one can predict memorability in novel sets of images for future (or already conducted) behavioral or neuroimaging studies. Such mem- orability prediction could also be valuable in a number of ap- plied settings (e.g., within education, marketing and human- computer interaction). Memorability researchers have employed computer vision models such as convolutional neural networks (CNNs) from early on (12), and advancements in the field have allowed researchers to predict image memorability with increasing precision (20–22). The inductive bias (the assumptions of the learning algorithms used to generalize to unseen data) of CNNs is inspired by knowledge about the primate visual sys- tem, and activations in the networks layers have, with some success, been used to explain neural activations (23). How- ever, some vulnerabilities of CNNs have been noted. For ex- ample, CNNs appear to depend more on image texture than biological vision systems do (24), and have problems with recognizing images based on the shape of objects (e.g., when texture is suppressed or removed). However, this vulnera- bility is reduced when the model’s shape bias is increased through training on shape representations (25). The LaMem train/test splits is a well-established benchmark for memorability prediction (12). The original MemNet (12), which is based on AlexNet (26), achieved a Spearman rank correlation of 0.64 on this benchmark. There have been several improvements on this benchmark, the leading ap- proaches utilize image captioning to enhance memorability predictions. That is, a CNN produces a textual description of the image, which is then used to provide more high-level se- mantic information which is embedded into a semantic vec- tor space before being combined with CNN image features in a multi-layered perceptron network. Squalli-Houssaini et al. (21) used this approach to reach a Spearman correlation of 0.72, with a mean squared error (MSE) of approximately 0.0092 (22). Leonardi et al. (22) used the captioning ap- proach with dual ResNet50s and a soft attention mechanism to reach a rank correlation of 0.687 with an MSE of 0.0079. The ResMem model (20), which is a CNN-based residual neural network architecture (ResNet), uses LaMem, but also takes advantage of a more recently published dataset named MemCat (11). This is a data set containing 10,000 images based on categories of animals, food, landscape, sports and vehicles. This data set also has a higher split half correla- tion than LaMem. Needell and Bainbridge (20) argue that the LaMem dataset on its own is lacking in generalizability due to poor sampling of naturalistic images. That is, the im- ages are more intended as artistic renderings designed to at- tract an online audience. Hence by combining MemCat with LaMem this should potentially yield a more generalizable model. Moreover, the increased size of the combined dataset might help in driving the model performance further than pre- vious models based on LaMem. The authors of ResMem also noted the importance of semantic information and struc- tured their approach to utilize semantic representations from a ResNet model in order to improve predictions. An added benefit of ResMem is that it is shared on the python pack- age index, which makes it easily accessible to researchers in diverse fields. Vision transformers. Vision transformers (ViT) have re- cently been shown to provide similar or better performance than CNNs in a variety of computer vision tasks (27). This architecture was first introduced in the natural language pro- cessing field (28) for capturing long-range dependencies in text. This architecture leads to superior speed/performance balance relativ to ResNet architectures (29). Moreover, ViTs have been shown to produce errors that are more similar to human errors (30), suggesting that they could take similar information into account (see also (31)). A reason for this may be that ViTs are likely to take more of the global context into account and be more dependent on the shape of objects rather than their texture (30). While it is not entirely clear why such properties may yield better predictions of image memorability, it could still help inform the discourse on the visual characteristics that are relevant as well as potentially yielding a better model for predicting image memorability. Hence, we set out to investigate if vision transformers can yield better predictions of memorability than the state-of- the-art in image memorability prediction. In particular, we aimed to (i) benchmark a model based on ViT against the well-established LaMem train/test splits (12), (ii) train a ViT against the combined LaMem and MemCat data sets (20) to benchmark against the ResMem model (20), (iii) train a final ViT model against a more diverse and deduplicated data set, (iv) validate the final ViT model against additional indepen- dent data sets and (v) inspect semantic level distributions of memorability scores for behavioral and predicted data. 2 | Hagen et al. | ViTMem Methods As our model is based on ViT to predict memorability we named it ViTMem. Because it has been shown that low- level visual features are less important for image memorabil- ity prediction, it would seem appropriate to use image aug- mentations in training our ViTMem model to reduce over- fitting. This approach have also been used by others (22), although not to the extent done here. The augmentations used consisted of horizontal flipping, sharpen, blur, motion blur, random contrast, hue saturation value, CLAHE, shift scale rotate, perspective, optical distortion and grid distortion (32). For training all models we used PyTorch, the ADAM optimizer and mean squared error (squared L2 norm) for the loss function. Images were input as batches of 32 in RGB and resized to 256x256 pixels before applying augmentations with a probability of 0.7 and center cropping to 224x224 pix- els. For creating ViTMem we used transfer learning on a vision transformer (27) model pretrained on ImageNet 1k (vit_base_patch16_224_miil) (33). The final classification layer was reduced to a single output and a sigmoid activation function. As we aim to provide an accessible model to the re- search community, it is also necessary to compare against the publicly available ResMem model. Unfortunately, the authors of ResMem did not publish their held-out test set, hence it is difficult to make a balanced compari- son between the currently published ResMem model and any competing models. We propose to do 10 train/test splits that can be used by future researchers (available at https://github.com/brainpriority/vitmem_data). Moreover, ResMem was not benchmarked on LaMem, hence a fair com- parison can only be made on the combined LaMem and MemCat data set. For the semantic level analysis, we chose to use image cap- tioning (34) as this provides an efficient method for deriv- ing semantic properties from images at scale. Importantly, as the image captioning model was trained on human image descriptions, it is likely to extract content that humans find meaningful in images, and in particular objects and contexts that are relevant for conveying such meanings. Hence, nouns derived from such descriptions are likely to be representative portions of the content that would convey meaning to humans observing the images. Data Sources. For the large-scale image memorability (LaMem) benchmark we used the LaMem dataset (12). The image set used by ResMem is a combination of the image sets LaMem (12) and MemCat (11). LaMem containing 58,741 and MemCat 10,000 images, for a total of 68,741 images. ResMem is reported to have used a held-out test set with 5000 images, hence we randomly selected 5000 images as our test set for our 10 train/test splits for this combined data set. For our final model we aimed to clean up the data and combine more of the available data sets on image memorability. As number of duplicated images within and between data sets is unknown and duplicated images may interfere with perfor- mance measures, we aimed to deduplicate the data for this model. Duplicated images were identified by simply deriv- ing embeddings from an off-the-shelf CNN model, and then visually inspecting the most similar embeddings. Our analy- sis of the data sets LaMem and MemCat showed that LaMem have 229 duplicated images while MemCat have 4. More- over, 295 of the images in LaMem is also in MemCat. We aimed to build a larger and more diverse data set by com- bining more sources, and for this we chose CVPR2011 (9) and FIGRIM (3). CVPR2011 had 6 internal duplicates, 651 duplicates against LaMem, 78 against MemCat og 9 against FIGRIM. FIGRIM had 20 duplicates against MemCat and 70 against LaMem. All identified duplicates were removed be- fore merging the data sets. As the images from FIGRIM and CVPR2011 were cropped, we obtained the original images before including them in the data set. This resulted in a data set with 71,658 images. For this data set we performed a 10% split for the test set. Results Results on LaMem data set. On the LaMem data set the ViTMem model reached an average Spearman rank correlation of 0.711 and an MSE of 0.0076 (see Table 1). Here we compare our performance to measures obtained by MemNet (12), Squalli-Houssaini et al. (21) and Leonardi et al. (22). Table 1. Comparison of model performance on LaMem data set Model MSE Loss ↓ Spearman ρ ↑ MemNet Unknown 0.640 Squalli-Houssaini et al. 0.0092 0.720 Leonardi et al. 0.0079 0.687 ViTMem 0.0076 0.711 Results on the combined LaMem and MemCat data set. Training on 10 train/test splits on the combined data set the results showed that ViTMem performed better than the ResMem model (see Table 2). The average across splits showed a Spearman rank correlation of 0.77 and an MSE of 0.005. Table 2. Model performance on LaMem and MemCat combiend dataset Model MSE Loss ↓ Spearman ρ ↑ ResMem 0.009 0.67 ViTMem 0.005 0.77 Results on combined and cleaned data set. To assess model performance on the larger and cleaned data set, we made a train/test split and then performed repeated k-fold cross validation with 10 train/test splits on the training set. This resulted in a mean MSE loss of 0.006 and a mean Spearman rank correlation of 0.76 (see Table 3). In order to provide a model for the community we used the full data Hagen et al. | ViTMem | 3 set to train the final model (ViTMem Final Model), which is published on the python package index as version 1.0.0. This was trained on the full training set and tested on its corresponding test set. The results showed a Spearman rank correlation of 0.77 and an MSE of 0.006 (see Table 3). The train/test splits are available on github. Table 3. Model performance on combined and cleaned data set Model MSE Loss ↓ Spearman ρ ↑ ViTMem 0.006 0.76 ViTMem Final Model 0.006 0.77 Validation on independent data sets. To further validate our model, we used memorability scores from an indepen- dent data set by Dubey and colleagues named PASCAL-S (7, 35) consisting of 852 images and cropped objects from the same images. ViTMem achieved a Spearman correlation of 0.44 on the images and 0.21 on the objects. In compar- ison ResMem achieved a correlation of 0.36 on the images and 0.14 on the objects. Validating against the THINGS data set (15), which consists of 26,106 images with memorabil- ity scores, achieved a Spearman rank correlation of 0.30 for ViTMem and 0.22 for ResMem. Semantic level analysis. In order to better understand how the model predictions relate to the semantic content of the images, we performed image captioning (34) on the com- bined LaMem and MemCat data set and the Places205 data set (36). We extracted nouns from the resulting image de- scriptions and averaged behavioral or predicted memorability scores for each noun (37). That is, the memorability for each image was assigned to each noun derived from the image cap- tioning procedure. For the combined LaMem and MemCat data set we averaged behavioral memorability scores over nouns (see Figure 1), while for the Places205 data set we averaged predicted memorability scores from the ViTMem model (see Figure 2). A general interpretation of the visu- alizations in Figure 1 and 2 is that they appear to reveal a dimension from nouns usually observed outdoors to more in- door related nouns and ending with nouns related to animals, and in particular, humans. This would appear to reflect the distributions observed in previous work (9, 15), and hence help to validate the model in terms of the image content it is sensitive to. To further investigate how well memorability associated with nouns were similar across the models we se- lected nouns occurring more than the 85th percentile in each set (654 nouns for LaMem and MemCat, 2179 nouns for Places205), this resulted in 633 matched nouns across sets. Analysis of these showed a Spearman ranked correlation of 0.89 and a R2 of 0.79, p<0.001 (see Figure 3). This analysis indicates that nouns from image captioning is a strong pre- dictor of image memorability and that the ViTMem model is able to generalize the importance of such aspects from the training set to a new set of images. Discussion Using vision transformers we have improved on the state- of-the-art in image memorability prediction. Results showed that ViTMem performed equal to or better than state-of- the-art models on LaMem, and better than ResMem on the LaMem and MemCat hybrid data set. In addition, we assem- bled a new deduplicated hybrid data set and benchmarked the ViTMem model against this before training a final model. The model was further validated on additional data sets, and performed better than ResMem on these as well. Finally, we ran a semantic level analysis by using image captioning on the hybrid data set. We ranked the behavioral memo- rability scores on the images, labeled with nouns extracted from the captioning procedure. The results revealed that im- ages labeled by nouns related to landscapes, cities, buildings and similar, were ranked lowest, whereas images labeled by nouns related to animate objects and food, were ranked high- est. This finding is consistent with known category effects on memorability (3, 7, 11, 12, 15) and suggests that the la- bels extracted from captioning procedure is strongly related to factors that drive memorability for those images. Subse- quently, we predicted memorability scores on images from a novel data set (Places205), ran the image captioning proce- dure, and ranked the predicted memorability scores on the images, labeled with nouns extracted from the captioning procedure. Visual inspection of the results revealed that the ranks were similar across samples and methods. This impres- sion was confirmed by a strong correlation between matching pairs of nouns and 79% explained variance, suggesting that ViTMem captures the semantic content that drives memora- bility in images. The use of image augmentations in training the ViTMem model in combination with state-of-the-art performance sug- gest that such augmentations are not disrupting the ability of the model to predict image memorability and hence may further support the importance of semantic level properties in image memorability. That is, the augmentations modify a range of low-level image properties but mostly leave the semantic content intact. In comparison with ResMem, which relies on a CNN-based residual neural network architecture, ViTMem is based on vision transformers which integrate information in a more global manner (30). As images are compositions of several semantically identifiable objects or parts of objects, a more holistic approach may be more apt at delineating the relative relevance of objects given their context. That is, we speculate that a broader integration of image features allows for a more complete evaluation of its constituent features in relation to each other. Hence, if semantic content is important for pre- dicting image memorability, the model may have weighed the importance of semantic components in relation to each other to a larger degree than models based on CNNs. ViTMem code and train/test sets are shared on github (https://github.com/brainpriority/), and a python package named vitmem is available on the python package index (see supplementary Sup. Note 1 for a tutorial). Researchers and interested parties can use the model to predict memorability 4 | Hagen et al. | ViTMem Memorability c() 0.56 0.58 0.60 0.62 0.64 0.66 0.68 0.70 0.72 0.74 0.76 0.78 0.80 0.82 0.84 0.86 0.88 0.90 mountains skyline clouds sunset fireplace view cloud dresser pine bedroom houses stream church highway waterfall house hotel sky boat wave water park reflection building walls tree people lights temple smoke flowers power rock group bus store lot truck fire center market game walking bench court person guitar police motorcycle food men show picture stem sign ground link women stuffed toy phone bride plate bag girl cards wedding shoe pair scarf hands hand shape neck face cut toothbrush half banana smile makeup necklace teeth pepper valley mountain dining lake trees buildings river hill city boats rocks fog lobby ocean tables middle kitchen area bridge construction field office room woods road clock photo steel street stove surfboard light dirt window side fence train bed museum door bird mirror flower grass course blue video row car couple table top line man bug case dog floor gas boy girls cell piece camera woman knife arms baby board gold head hair sunglasses persons shirt tie feet nose chocolate word beard snake tattoo blood bikini Fig. 1. Average behavioral image memorability scores for nouns that were extracted from images in the LaMem and MemCat data sets. The nouns shown are those that occurred most frequently or that are more frequent in the English language (38). Memorability c() 0.52 0.54 0.56 0.58 0.60 0.62 0.64 0.66 0.68 0.70 0.72 0.74 0.76 0.78 0.80 0.82 0.84 0.86 0.88 0.90 badlands rim stormy glacier mountain sun town hill fireplace houses sunset snow slope desert dusk rocks couches city cabinets steeple university street place hotel highway cathedral formation center building beach tree home way stone lot fire christmas lighthouse space monument desk people crowd boat wall inside airport music van museum model round stage statue baseball auditorium party classroom tent stand row court store picture pink bus shelf bowling sale men bars family gym fish boy motel woman ring rack soldier girl ties dresses words name dancing suit wrestlers arms mannequin cookies cream shirt wife cupcakes chocolate bikini hillside clouds mountains valley cloud farm village snowy square waves vineyard island view mansion smoke castle living coast lawn area church house tower clock field road rain wave sink top state water chairs bed room side birds dock leaves park supplies force station table play post cross market desks photos group image library game line school video dog food star crib show clothes book floor children man heart baby display sign roller women class football girls case hands team desserts face shirts suits logo hair plate pastries head grave meat tie bread donuts mouth dance dress dancer Fig. 2. Average ViTMem predicted image memorability scores for nouns that were extracted from images in the Places205 data set. The nouns shown are those that occurred most frequently or that are more frequent in the English language (38). 0.6 0.7 0.8 0.9 0.6 0.7 0.8 0.9 Memorability for LaMem & MemCat Nouns (Behavioral) Memorability for Places205 Nouns (ViTMem) Fig. 3. Average memorability scores for images with matching nouns in different data sets. The y-axis shows average predicted memorability scores from ViTMem on the Places205 data set. The x-axis shows average behavioral memorability scores on the combined LaMem and MemCat data set. in existing or novel stimuli and employ them in research or applied settings. The ViTMem model will allow researchers to more precisely predict image memorability. The release of ViTMem follows up ResMem in providing an accessible method for predicting image memorability. This is impor- tant for studies aiming to control for how easily an image can be remembered. This will for example allow experimental psychologists and neuroscientists to better control their re- search. Similarly, educators, advertisers and visual designers can leverage the model to improve the memorability of their content. Despite state-of-the-art performance in memorability predic- tion, improvements may still be possible to achieve. Previous works have shown benefits of pretraining their networks on data sets of places and objects prior to fine tuning for memo- rability prediction (39). Moreover, ViTMem do not take im- age captioning into account, which have been successfully done with CNNs (21, 22). Hence there is potentially more to be gained from incorporating image semantics and/or pre- training on data sets of objects and places. In addition, ViT- Mem is only based on the "base" configuration of the avail- able ViT models. Model performance may still increase by adopting the “large” or “huge” configurations of the model. We conclude that ViTMem can be used to predict memora- bility for images at a level that is equal to or better than state- of-the-art models, and we propose that vision transformers provide a new step forward in the computational prediction of image memorability. Hagen et al. | ViTMem | 5 References 1. Phillip Isola, Jianxiong Xiao, Devi Parikh, Antonio Torralba, and Aude Oliva. What makes a photograph memorable? IEEE transactions on pattern analysis and machine intelligence, 36(7):1469–1482, 2013. 2. Lore Goetschalckx, Pieter Moors, and Johan Wagemans. Image memorability across longer time intervals. Memory, 26(5):581–588, 2018. 3. Zoya Bylinskii, Phillip Isola, Constance Bainbridge, Antonio Torralba, and Aude Oliva. In- trinsic and extrinsic effects on image memorability. Vision research, 116:165–178, 2015. 4. Lore Goetschalckx, Jade Moors, and Johan Wagemans. Incidental image memorability. Memory, 27(9):1273–1282, 2019. 5. Zoya Bylinskii, Lore Goetschalckx, Anelise Newman, and Aude Oliva. Memorability: An image-computable measure of information utility. In Human Perception of Visual Informa- tion, pages 207–239. Springer, 2022. 6. Nicole C Rust and Vahid Mehrpour. Understanding image memorability. Trends in cognitive sciences, 24(7):557–568, 2020. 7. Rachit Dubey, Joshua Peterson, Aditya Khosla, Ming-Hsuan Yang, and Bernard Ghanem. What makes an object memorable? In Proceedings of the ieee international conference on computer vision, pages 1089–1097, 2015. 8. Wilma A Bainbridge, Daniel D Dilks, and Aude Oliva. Memorability: A stimulus-driven per- ceptual neural signature distinctive from memory. NeuroImage, 149:141–152, 2017. 9. Phillip Isola, Devi Parikh, Antonio Torralba, and Aude Oliva. Understanding the intrinsic memorability of images. Advances in neural information processing systems, 24, 2011. 10. Timothy F Brady, Talia Konkle, and George A Alvarez. A review of visual memory capacity: Beyond individual items and toward structured representations. Journal of vision, 11(5): 4–4, 2011. 11. Lore Goetschalckx and Johan Wagemans. Memcat: a new category-based image set quan- tified on memorability. PeerJ, 7:e8169, 2019. 12. Aditya Khosla, Akhil S. Raju, Antonio Torralba, and Aude Oliva. Understanding and predict- ing image memorability at a large scale. In International Conference on Computer Vision (ICCV), 2015. 13. Michael Gygli, Helmut Grabner, Hayko Riemenschneider, Fabian Nater, and Luc Van Gool. The interestingness of images. In Proceedings of the IEEE international conference on computer vision, pages 1633–1640, 2013. 14. Wilma A Bainbridge. The resiliency of image memorability: A predictor of memory separate from attention and priming. Neuropsychologia, 141:107408, 2020. 15. Max A. Kramer, Martin N. Hebart, Chris I. Baker, and Wilma A. Bainbridge. The features underlying the memorability of objects. bioRxiv, 2022. doi: 10.1101/2022.04.29.490104. 16. Eleanor Rosch, Carolyn B Mervis, Wayne D Gray, David M Johnson, and Penny Boyes- Braem. Basic objects in natural categories. Cognitive psychology, 8(3):382–439, 1976. 17. Douglas L Medin and John D Coley. Concepts and categorization. Perception and cognition at century’s end: Handbook of perception and cognition, pages 403–439, 1998. 18. Erdem Akagunduz, Adrian G Bors, and Karla K Evans. Defining image memorability using the visual memory schema. IEEE transactions on pattern analysis and machine intelligence, 42(9):2165–2178, 2019. 19. Muxuan Lyu, Kyoung Whan Choe, Omid Kardan, Hiroki P Kotabe, John M Henderson, and Marc G Berman. Overt attentional correlates of memorability of scene images and their relationships to scene semantics. Journal of Vision, 20(9):2–2, 2020. 20. Coen D Needell and Wilma A Bainbridge. Embracing new techniques in deep learning for estimating image memorability. Computational Brain & Behavior, pages 1–17, 2022. 21. Hammad Squalli-Houssaini, Ngoc QK Duong, Marquant Gwenaëlle, and Claire-Hélène De- marty. Deep learning for predicting image memorability. In 2018 IEEE international con- ference on acoustics, speech and signal processing (ICASSP), pages 2371–2375. IEEE, 2018. 22. Marco Leonardi, Luigi Celona, Paolo Napoletano, Simone Bianco, Raimondo Schettini, Franco Manessi, and Alessandro Rozza. Image memorability using diverse visual features and soft attention. In International Conference on Image Analysis and Processing, pages 171–180. Springer, 2019. 23. Daniel LK Yamins, Ha Hong, Charles F Cadieu, Ethan A Solomon, Darren Seibert, and James J DiCarlo. Performance-optimized hierarchical models predict neural responses in higher visual cortex. Proceedings of the national academy of sciences, 111(23):8619–8624, 2014. 24. Nicholas Baker, Hongjing Lu, Gennady Erlikhman, and Philip J Kellman. Local features and global shape information in object classification by deep convolutional neural networks. Vision research, 172:46–61, 2020. 25. Robert Geirhos, Patricia Rubisch, Claudio Michaelis, Matthias Bethge, Felix A Wichmann, and Wieland Brendel. Imagenet-trained cnns are biased towards texture; increasing shape bias improves accuracy and robustness. arXiv preprint arXiv:1811.12231, 2018. 26. Alex Krizhevsky, Ilya Sutskever, and Geoffrey E Hinton. Imagenet classification with deep convolutional neural networks. Advances in neural information processing systems, 25, 2012. 27. Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv preprint arXiv:2010.11929, 2020. 28. Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. Attention is all you need. Advances in neural information processing systems, 30, 2017. 29. Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Identity mappings in deep residual networks. In European conference on computer vision, pages 630–645. Springer, 2016. 30. Shikhar Tuli, Ishita Dasgupta, Erin Grant, and Thomas L Griffiths. Are convolutional neural networks or transformers more like human vision? arXiv preprint arXiv:2105.07197, 2021. 31. Nicholas Baker and James H Elder. Deep learning models fail to capture the configural nature of human shape perception. Iscience, 25(9):104913, 2022. 32. Alexander Buslaev, Vladimir I Iglovikov, Eugene Khvedchenya, Alex Parinov, Mikhail Druzhinin, and Alexandr A Kalinin. Albumentations: fast and flexible image augmentations. Information, 11(2):125, 2020. 33. Tal Ridnik, Emanuel Ben-Baruch, Asaf Noy, and Lihi Zelnik-Manor. Imagenet-21k pretrain- ing for the masses. arXiv preprint arXiv:2104.10972, 2021. 34. Peng Wang, An Yang, Rui Men, Junyang Lin, Shuai Bai, Zhikang Li, Jianxin Ma, Chang Zhou, Jingren Zhou, and Hongxia Yang. Unifying architectures, tasks, and modalities through a simple sequence-to-sequence learning framework. arXiv preprint arXiv:2202.03052, 2022. 35. Yin Li, Xiaodi Hou, Christof Koch, James M Rehg, and Alan L Yuille. The secrets of salient object segmentation. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 280–287, 2014. 36. Bolei Zhou, Agata Lapedriza, Jianxiong Xiao, Antonio Torralba, and Aude Oliva. Learning deep features for scene recognition using places database. Advances in neural information processing systems, 27, 2014. 37. Steven Loria et al. textblob v0.17.1, October 2021. 38. Robyn Speer. rspeer/wordfreq: v3.0, September 2022. 39. Shay Perera, Ayellet Tal, and Lihi Zelnik-Manor. Is image memorability prediction solved? In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition workshops, pages 0–0, 2019. 6 | Hagen et al. | ViTMem Supplementary Note 1: How to use the vitmem python package Python needs to be installed on a computer before pip can be used to install the vitmem package. To install vitmem from a command prompt run: pip install vitmem To predict image memorability for an image named "image.jpg", run the following in a python interpreter: from vitmem import ViTMem model = ViTMem() memorability = model("image.jpg") print(f"Predicted memorability: {memorability}") Hagen et al. | ViTMem | 7 |