-
Notifications
You must be signed in to change notification settings - Fork 20
Description
Hi,
I am running GCond on the Amazon dataset. The setting is as follows: I have added appropriate data loading, train-val-test split creation script in your code. And now I run it to generate say 73 nodes out of ~7000 in Photos dataset, or ~130 nodes out of ~13000 nodes of Computers dataset (around 1% of the size). Then, I use these datasets to train a simple GCN (2-layered, with ReLU between the two layers, no batchnorm, no dropout, optimized with standard parameters of Adam, the standard stuff). This model fails to perform well. This model only predicts one single class. The performance is ~24% for Photos and ~37% for Computers, while the full train set performance is ~90% for both. And random gets to about ~66-72% for both.
So, is there a quick fix, like finding the optimum inner-outer loop parameters, or optimizing the SGC parameters?
P.S.: I noticed that the loss (gradient matching loss) printed out by your code at the 0th epoch is too high for these datasets (~97) which doesn't go down a lot, while for datasets like Cora, Citeseer, etc. it starts with ~1-3 and falls down below 1 very quickly.