-
Notifications
You must be signed in to change notification settings - Fork 20
add visual search #51
base: master
Are you sure you want to change the base?
Conversation
vs_model_param = {} | ||
vs_model_param['tar_pool'] = target_model_pool | ||
vs_model_param['stim_pool'] = stimuli_model_pool | ||
vs_model_param['model_layers'] = visual_search_layer | ||
vs_model_param['tar_size'] = target_img_size | ||
vs_model_param['stim_size'] = stimuli_img_size |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
also confused by these here
layers = { | ||
'vgg-16': [f'block{i + 1}_pool' for i in range(3,5)], | ||
} |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I would create another vgg-16 BrainModel from the same underlying BaseModel that makes these exact commitments
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
The issue is that the "target_image" size is just 28x28 pixels. So we will need to exclusively define which layer will work for a specific ML model. So isn't it much better that we commit these layers in a separate py file? i.e. which layer to use for which ML model?
check score_model example on visual search benchmark.