-
Notifications
You must be signed in to change notification settings - Fork 2
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Graph generative models #26
Comments
I am quoting what I wrote last in #24 :
|
Is this something that maybe @dnguyen1196 would like to take point on? To really dig down on this specific aspect? |
I totally agree that this is a very interesting area. But I'm not sure if I understand how does this fit in our current project at this stage. (sure once we move on to RL regime we need some machinery to navigate the discrete chemical space) |
Also, structuring |
Graph generative models are the part that does the unsupervised or self-supervised training for graph structures, as I mention above. You already tried playing with an instance of such models, but have not gotten this to work well. I am pitching to focus a specific project which needs someone to focus on this specific topic here, as this is more important/complementary in its benefits than being Bayesian about parts of the model since it is not an inference issue, but a modeling issue that unlocks the ability to use more data than we have measurements for to improve a representation. Remember, the Cambridge group got a lot of mileage out of doing this and this is already at the heart of our story as mentioned in issue #3 This is not just necessary for RL, it is necessary for all things we may want to do if we want to train graph-representations on more data than we have measurements for (but have graphs). |
I am arguing to start focusing these discussions about graph-space here in a specific issue instead of throwing random self-supervised models or eigenspaces of graphs into the loop in other issues. |
So this would be a kind of regularized representation of graphs on fixed-dimensional spaces. And we might not need the actual generative part for now? |
If you see my starting comments, I care mostly about the joint space of models that do unsupervised and self-supervised graph representation learning and to compare and evaluate them specifically. Whether that is actually a graph VAE I do not care, it can be a self-supervised thing. But we need a space of graph representations that we can train from graphs alone and potentially use within a semi-supervised framework and it is prudent to have an issue and a focused effort targeted at exploring that systematically and focusing on experiments and literature review to create clarity about what works. |
Ah okay I see your point. To start the discussion, I think it might be helpful to list a few flavors of graph generative models.
|
Could you also please add titles so we can quickly parse what this work is and add a category with the self-supervised space of work? I. e. I envision: Graph VAEs:
Self-supervised work: Comparisons and performance... I also added an introductory blog for those not familiar with all these concepts above and a recent paper I saw targeted at molecular stuff. And I will inquire again: |
I think I can take a stab at this since I've played with these before. It's just that for some of these methods here (especially the ones that encode graph as a sequence of actions) might take some decent time to make them work. I'd suggest lower the priority of testing those. |
I know you have played with this before, but I am arguing for a full-time effort just on this narrow topic by somebody that would slot into the overall project in order to make more measurable progress with empirical evidence. Do you want to focus on this for a few weeks? |
How about, I'll aim to have the full characterization results of the simpler models here before our discussion on Fri. After that we can plan accordingly if we decided to also have more sophisticated algorithms. |
Sounds reasonable. |
I realized that Kipf and Welling's variational graph autoencoder VGAE https://arxiv.org/abs/1611.07308 and this paragraph-vector based semi-supervised model https://arxiv.org/pdf/1711.10168.pdf might be the same thing with particular choices of pooling functions and training schedule. |
Graph generative models are important for the tasks we have been describing.
The core idea is to posit a model which defines some distribution over graphs
P(G)
, for instance via a low dimensional modelP(Z)
.Example:
P(G,Z) = P(G|Z) P(Z)
like the class of models commonly described as Graph VAEs.This is the probabilistic modeling of graphs space, there is also the self-supervised world which seems to implicitly represent
P(G)
without as clear a generative story.There are also pure deep learning things, like transformer-type models and autoregressive graph models.
I am unsure how those would compare empirically in terms of performance.
In this issue, I suggest we survey the landscape of these models and their empirical comparisons and sketch out a strategy to compare them.
I referenced some work in a previous issue #24 , I pitch that we move this discussion here.
The text was updated successfully, but these errors were encountered: