This project focuses on implementing FastGCN, a scalable alternative to traditional Graph Convolutional Networks (GCNs) that leverages importance sampling to improve efficiency. Additionally, we explore adaptive sampling techniques to further enhance accuracy and computational performance.
- Pooya Nasiri ([email protected])
- Giacomo Vettoretti ([email protected])
- Jacopo Righetto ([email protected])
- Implementation of FastGCN with importance sampling
- Adaptive sampling strategy to improve accuracy
- Comparative analysis with traditional GCNs
- Experiments conducted on benchmark datasets: MNIST, Cora, PubMed, CiteSeer
- Performance evaluation based on training time, accuracy, and memory usage
To run the project, install the necessary dependencies:
pip install torch torchvision numpy scipy scikit-learn matplotlibThe project is implemented in Jupyter Notebook. To run the notebook:
jupyter notebook FastGCN.ipynb- Load the dataset (MNIST, Cora, PubMed, or CiteSeer)
- Preprocess the graph data (adjacency matrix, feature normalization)
- Train FastGCN using the implemented model
- Evaluate results using classification accuracy and computational efficiency
- MNIST: Converted into a graph format where nodes represent images.
- Cora: Citation network of scientific publications categorized into seven classes.
- PubMed: Biomedical articles classified into three categories.
- CiteSeer: Research paper citation network with six categories.
- FastGCN reduces training time compared to traditional GCNs.
- Adaptive sampling improves accuracy by selecting influential nodes dynamically.
- Trade-off observed between efficiency and accuracy, with importance sampling optimizing computational cost.
- Dataset-specific tuning is crucial for achieving optimal performance.
| Dataset | Test Accuracy |
|---|---|
| MNIST | 0.2500 |
| PubMed | 0.5000 |
| Cora | 0.1407 |
| CiteSeer | 0.2218 |
- Exploring hybrid sampling techniques combining multiple strategies.
- Fine-tuning importance sampling distributions for different datasets.
- Implementing hierarchical sampling for large-scale graphs.
- Integrating adaptive learning rates for better convergence.
- Jie Chen, Tengfei Ma, and Cao Xiao. FastGCN: Fast learning with graph convolutional networks via importance sampling (ICLR, 2018).
- William L. Hamilton, Rex Ying, and Jure Leskovec. Inductive representation learning on large graphs (2018).
- Thomas N. Kipf and Max Welling. Semi-supervised classification with graph convolutional networks (2017).