Sample stimuli

sample 0 sample 1 sample 2 sample 3 sample 4 sample 5 sample 6 sample 7 sample 8 sample 9

How to use

from brainscore_vision import load_benchmark
benchmark = load_benchmark("Baker2022frankenstein-accuracy_delta")
score = benchmark(my_model)

Model scores

Min Alignment Max Alignment

Rank

Model

Score

1
.983
2
.974
3
.972
4
.939
5
.929
6
.929
7
.929
8
.899
9
.897
10
.889
11
.887
12
.874
13
.873
14
.867
15
.867
16
.865
17
.862
18
.858
19
.855
20
.854
21
.850
22
.844
23
.824
24
.806
25
.803
26
.801
27
.798
28
.793
29
.775
30
.770
31
.767
32
.767
33
.766
34
.766
35
.761
36
.754
37
.741
38
.731
39
.731
40
.727
41
.724
42
.717
43
.715
44
.714
45
.713
46
.712
47
.698
48
.696
49
.692
50
.681
51
.680
52
.671
53
.666
54
.662
55
.658
56
.645
57
.644
58
.641
59
.623
60
.617
61
.608
62
.607
63
.599
64
.578
65
.572
66
.568
67
.568
68
.567
69
.558
70
.545
71
.539
72
.536
73
.527
74
.523
75
.503
76
.503
77
.494
78
.487
79
.485
80
.485
81
.465
82
.464
83
.463
84
.463
85
.462
86
.457
87
.454
88
.448
89
.448
90
.448
91
.435
92
.427
93
.420
94
.418
95
.410
96
.398
97
.395
98
.378
99
.374
100
.372
101
.365
102
.365
103
.362
104
.343
105
.339
106
.335
107
.335
108
.335
109
.334
110
.314
111
.313
112
.310
113
.308
114
.301
115
.288
116
.286
117
.284
118
.282
119
.278
120
.276
121
.270
122
.257
123
.232
124
.232
125
.220
126
.212
127
.204
128
.201
129
.160
130
.156
131
.155
132
.142
133
.135
134
.131
135
.119
136
.111
137
.096
138
.089
139
.042
140
.038
141
.007
142
.006
143
.003
144
.002
145
.000
146
.000
147
.000
148
.000
149
.000
150
.000
151
.000
152
.000
153
.000
154
.000
155
.000
156
.000
157
.000
158
.000
159
.000
160
.000
161
.000
162
.000
163
.000
164
.000
165
.000
166
.000
167
.000
168
.000
169
.000
170
.000
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238

Benchmark bibtex

@article{BAKER2022104913,
                title = {Deep learning models fail to capture the configural nature of human shape perception},
                journal = {iScience},
                volume = {25},
                number = {9},
                pages = {104913},
                year = {2022},
                issn = {2589-0042},
                doi = {https://doi.org/10.1016/j.isci.2022.104913},
                url = {https://www.sciencedirect.com/science/article/pii/S2589004222011853},
                author = {Nicholas Baker and James H. Elder},
                keywords = {Biological sciences, Neuroscience, Sensory neuroscience},
                abstract = {Summary
                A hallmark of human object perception is sensitivity to the holistic configuration of the local shape features of an object. Deep convolutional neural networks (DCNNs) are currently the dominant models for object recognition processing in the visual cortex, but do they capture this configural sensitivity? To answer this question, we employed a dataset of animal silhouettes and created a variant of this dataset that disrupts the configuration of each object while preserving local features. While human performance was impacted by this manipulation, DCNN performance was not, indicating insensitivity to object configuration. Modifications to training and architecture to make networks more brain-like did not lead to configural processing, and none of the networks were able to accurately predict trial-by-trial human object judgements. We speculate that to match human configural sensitivity, networks must be trained to solve a broader range of object tasks beyond category recognition.}
        }

Ceiling

Not available

Data: Baker2022frankenstein

Metric: accuracy_delta