Reproducing, breaking, and iterating Graph Metanetworks for Proessing Diverse Neural Architectures
Creating graph representations of neural networks, called parameter graphs. We then train a graph neural network to make predictions about neural architectures and their weight spaces based on their parameter graph.
- To scale graph metanetworks to larger and deeper models, facilitating tasks such as architecture search.
- To apply similar metanetworks to make predictions over activation spaces, not just weight spaces.
- To explore the application of metanets to dictionary learning and compare their performance against sparse autoencoders.
Currently, we can successfully represent linear, convolutional, normalization, and activation layers in parameter graphs. We are in the process of generating the training data and developing the graph neural network to perform meta predictions.
Simply clone the open graph_playground.ipynb to generate and visiualize parameter graphs from a neural network.
Interested? Join our research updates list!
Stay updated on our progress and get a behind-the-scenes look at our ambitious projects.