Sample stimuli

sample 0 sample 1 sample 2 sample 3 sample 4 sample 5 sample 6 sample 7 sample 8 sample 9

How to use

from brainscore_vision import load_benchmark
benchmark = load_benchmark("MajajHong2015.V4-pls")
score = benchmark(my_model)

Model scores

Min Alignment Max Alignment

Rank

Model

Score

1
.777
2
.750
3
.714
4
.698
5
.663
6
.655
7
.624
8
.620
9
.614
10
.614
11
.611
12
.610
13
.610
14
.610
15
.610
16
.608
17
.607
18
.605
19
.605
20
.604
21
.604
22
.604
23
.603
24
.603
25
.602
26
.602
27
.602
28
.602
29
.602
30
.602
31
.602
32
.602
33
.602
34
.602
35
.602
36
.602
37
.602
38
.602
39
.601
40
.601
41
.601
42
.600
43
.600
44
.600
45
.600
46
.600
47
.599
48
.599
49
.599
50
.599
51
.599
52
.599
53
.598
54
.598
55
.598
56
.598
57
.598
58
.598
59
.597
60
.597
61
.597
62
.597
63
.596
64
.596
65
.596
66
.596
67
.596
68
.596
69
.596
70
.595
71
.595
72
.595
73
.594
74
.593
75
.592
76
.592
77
.592
78
.592
79
.592
80
.592
81
.592
82
.592
83
.592
84
.592
85
.591
86
.591
87
.591
88
.591
89
.591
90
.591
91
.591
92
.590
93
.590
94
.590
95
.590
96
.589
97
.589
98
.589
99
.589
100
.589
101
.589
102
.589
103
.589
104
.589
105
.589
106
.588
107
.588
108
.588
109
.588
110
.588
111
.588
112
.587
113
.587
114
.587
115
.587
116
.587
117
.586
118
.586
119
.586
120
.586
121
.586
122
.585
123
.585
124
.585
125
.585
126
.584
127
.584
128
.584
129
.584
130
.584
131
.584
132
.584
133
.584
134
.583
135
.583
136
.583
137
.583
138
.583
139
.583
140
.583
141
.582
142
.582
143
.582
144
.582
145
.582
146
.582
147
.582
148
.582
149
.582
150
.582
151
.581
152
.581
153
.581
154
.581
155
.581
156
.581
157
.581
158
.580
159
.580
160
.580
161
.580
162
.580
163
.580
164
.580
165
.580
166
.579
167
.579
168
.579
169
.579
170
.579
171
.579
172
.578
173
.578
174
.578
175
.578
176
.578
177
.578
178
.578
179
.578
180
.578
181
.577
182
.577
183
.577
184
.577
185
.577
186
.576
187
.576
188
.576
189
.575
190
.575
191
.575
192
.575
193
.575
194
.575
195
.574
196
.574
197
.574
198
.574
199
.574
200
.574
201
.574
202
.574
203
.574
204
.574
205
.574
206
.573
207
.573
208
.573
209
.573
210
.573
211
.572
212
.572
213
.572
214
.571
215
.571
216
.571
217
.571
218
.570
219
.570
220
.570
221
.570
222
.570
223
.570
224
.570
225
.570
226
.570
227
.569
228
.569
229
.569
230
.569
231
.569
232
.569
233
.569
234
.569
235
.569
236
.569
237
.568
238
.568
239
.568
240
.568
241
.568
242
.568
243
.568
244
.567
245
.567
246
.567
247
.566
248
.566
249
.566
250
.566
251
.566
252
.566
253
.566
254
.566
255
.565
256
.565
257
.565
258
.564
259
.564
260
.564
261
.564
262
.563
263
.563
264
.563
265
.563
266
.562
267
.562
268
.562
269
.562
270
.562
271
.562
272
.561
273
.560
274
.560
275
.560
276
.560
277
.560
278
.560
279
.560
280
.559
281
.559
282
.558
283
.558
284
.558
285
.558
286
.558
287
.558
288
.558
289
.557
290
.557
291
.557
292
.557
293
.556
294
.556
295
.556
296
.555
297
.555
298
.555
299
.555
300
.555
301
.554
302
.554
303
.553
304
.553
305
.553
306
.551
307
.551
308
.551
309
.550
310
.550
311
.550
312
.550
313
.550
314
.550
315
.549
316
.549
317
.549
318
.549
319
.549
320
.549
321
.548
322
.548
323
.548
324
.548
325
.548
326
.548
327
.547
328
.547
329
.546
330
.545
331
.545
332
.544
333
.544
334
.543
335
.542
336
.541
337
.541
338
.541
339
.540
340
.539
341
.539
342
.539
343
.538
344
.538
345
.537
346
.537
347
.536
348
.536
349
.536
350
.533
351
.533
352
.533
353
.531
354
.530
355
.530
356
.527
357
.526
358
.524
359
.523
360
.521
361
.519
362
.518
363
.517
364
.517
365
.516
366
.516
367
.516
368
.515
369
.515
370
.514
371
.514
372
.514
373
.514
374
.514
375
.514
376
.514
377
.514
378
.514
379
.514
380
.514
381
.513
382
.511
383
.511
384
.509
385
.509
386
.504
387
.504
388
.503
389
.501
390
.501
391
.498
392
.498
393
.497
394
.494
395
.494
396
.491
397
.489
398
.487
399
.487
400
.486
401
.485
402
.485
403
.483
404
.481
405
.476
406
.473
407
.472
408
.469
409
.466
410
.462
411
.456
412
.454
413
.453
414
.452
415
.451
416
.445
417
.443
418
.439
419
.438
420
.437
421
.437
422
.436
423
.433
424
.433
425
.432
426
.432
427
.431
428
.431
429
.431
430
.430
431
.430
432
.427
433
.421
434
.420
435
.419
436
.418
437
.409
438
.376
439
.376
440
.342
441
.339
442
.328
443
.316
444
.185
445
.179
446
.154
447
.098
448
.078
449
.073
450
.068
451
.068
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470

Benchmark bibtex

@article {Majaj13402,
            author = {Majaj, Najib J. and Hong, Ha and Solomon, Ethan A. and DiCarlo, James J.},
            title = {Simple Learned Weighted Sums of Inferior Temporal Neuronal Firing Rates Accurately Predict Human Core Object Recognition Performance},
            volume = {35},
            number = {39},
            pages = {13402--13418},
            year = {2015},
            doi = {10.1523/JNEUROSCI.5181-14.2015},
            publisher = {Society for Neuroscience},
            abstract = {To go beyond qualitative models of the biological substrate of object recognition, we ask: can a single ventral stream neuronal linking hypothesis quantitatively account for core object recognition performance over a broad range of tasks? We measured human performance in 64 object recognition tests using thousands of challenging images that explore shape similarity and identity preserving object variation. We then used multielectrode arrays to measure neuronal population responses to those same images in visual areas V4 and inferior temporal (IT) cortex of monkeys and simulated V1 population responses. We tested leading candidate linking hypotheses and control hypotheses, each postulating how ventral stream neuronal responses underlie object recognition behavior. Specifically, for each hypothesis, we computed the predicted performance on the 64 tests and compared it with the measured pattern of human performance. All tested hypotheses based on low- and mid-level visually evoked activity (pixels, V1, and V4) were very poor predictors of the human behavioral pattern. However, simple learned weighted sums of distributed average IT firing rates exactly predicted the behavioral pattern. More elaborate linking hypotheses relying on IT trial-by-trial correlational structure, finer IT temporal codes, or ones that strictly respect the known spatial substructures of IT ({	extquotedblleft}face patches{	extquotedblright}) did not improve predictive power. Although these results do not reject those more elaborate hypotheses, they suggest a simple, sufficient quantitative model: each object recognition task is learned from the spatially distributed mean firing rates (100 ms) of \~{}60,000 IT neurons and is executed as a simple weighted sum of those firing rates.SIGNIFICANCE STATEMENT We sought to go beyond qualitative models of visual object recognition and determine whether a single neuronal linking hypothesis can quantitatively account for core object recognition behavior. To achieve this, we designed a database of images for evaluating object recognition performance. We used multielectrode arrays to characterize hundreds of neurons in the visual ventral stream of nonhuman primates and measured the object recognition performance of \>100 human observers. Remarkably, we found that simple learned weighted sums of firing rates of neurons in monkey inferior temporal (IT) cortex accurately predicted human performance. Although previous work led us to expect that IT would outperform V4, we were surprised by the quantitative precision with which simple IT-based linking hypotheses accounted for human behavior.},
            issn = {0270-6474},
            URL = {https://www.jneurosci.org/content/35/39/13402},
            eprint = {https://www.jneurosci.org/content/35/39/13402.full.pdf},
            journal = {Journal of Neuroscience}}

Ceiling

0.90.

Note that scores are relative to this ceiling.

Data: MajajHong2015.V4

2560 stimuli recordings from 88 sites in V4

Metric: pls