You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
We're interested in checking what the object learner does when it decodes unknown objects. We want to run a "rehearsal" experiment here.
As part of this task, you'll need to:
Run the unknown objects script on the M5 objects train curriculum, removing the list of objects from earlier (from your look at object contrasts and the discussion). That is:
Evaluate that GNN on train and on test. It's worth recording train accuracy for completeness though it doesn't tell us much. The main reason to run inference on train is to produce the decode files ADAM needs (the feature.yaml s). Also record the test accuracy. I'd expect this to be very similar to ADAM's test accuracy.
Run ADAM with the subset learner over the resulting curriculum / decode. Parameters would be similar to the M5 objects curriculum except for the train curriculum.
Since it's ~easy to do (we already have observers set up to do it), it's probably worth collecting/comparing per-object accuracy and qualitative outcomes results vs. the m5_objects_v0_with_mugs baseline results.
The main interesting thing to look at here are the per-object results. I think this means:
A table of ADAM's train vs. test accuracies in the baseline (m5_objects_v0_with_mugs) case and for "unknown objects." (Mostly for completeness.)
The confusion matrices for ADAM's output on train and on test. (details on confusion matrices: see Downsampling experiment #1156.)
Maybe separate matrices for the GNN?
Separately, a plot of only those rows of the test-time confusion matrix (or matrices) that shows only the objects removed from train (the "unknown objects").
A writeup containing these figures plus discussion focusing on the unknown objects outcomes.
e.g. we expect the GNN to confuse unknown objects with ones it saw during training, how weird vs. "reasonable" are the confusions it makes. Is it confusing apple/ball with orange, or with book? etc.
This probably requires manually looking at the stroke and stroke graph images for the unknown objects in the test set.
The text was updated successfully, but these errors were encountered:
We're interested in checking what the object learner does when it decodes unknown objects. We want to run a "rehearsal" experiment here.
As part of this task, you'll need to:
feature.yaml
s). Also record the test accuracy. I'd expect this to be very similar to ADAM's test accuracy.m5_objects_v0_with_mugs
baseline results.m5_objects_v0_with_mugs
) case and for "unknown objects." (Mostly for completeness.)The text was updated successfully, but these errors were encountered: