Boosting Graph Neural Networks with getML: Automated Feature Engineering for Superior Model Performance
How getML’s FastProp algorithm helps optimize your Graph Neural Network
Graph Neural Networks (GNNs) excel at modeling complex data but do require feature engineering. This article shows how integrating getML's FastProp automates this process, boosting accuracy to 92.5% on the CORA dataset—surpassing the 90.16% benchmark. This approach simplifies implementation, reduces manual effort, and ensures consistent performance across a wide range of neural architectures, making it a valuable tool for data scientists for building consistent and high-performing models with minimal effort.
Introduction
While Large Language models and Diffusion models have been attracting a lot of attention, Graph Neural Networks (GNN) have also advanced in leaps and bounds. No longer confined to academic exercises, GNNs are applied in recommender systems, hardware development, drug discovery, and much more.
A key advantage of Graph Neural Networks (GNNs) lies in their ability to natively process relational data, making them suited for tasks like node and link prediction. Unlike traditional deep learning architectures, such as convolutional or recurrent neural networks, which require Euclidean data representations, GNNs operate directly on graph structures. This eliminates the need for extensive feature engineering — transforming relational data into uniform feature vectors — a prerequisite for other architectures. This is particularly valuable, as feature engineering is a time-consuming and cumbersome process that often resembles an art more than a technique, heavily relying on domain knowledge and intuition.
While an abundance of options and parameters when working with GNNs may excite experts, it presents a considerable challenge to novices and casual practitioners: They require extensive domain knowledge, careful graph preprocessing, and significant computational resources.
Another potentially lucrative but labor-intensive and computationally expensive step is the experimentation with a variety of Neural Network architectures.
Thus, to fully harness the power of GNNs,
- a lot of time consuming manual work and
- a substantial computation budget
is required.
But not anymore! This article introduces an approach that not only eliminates the need for intricate, manual feature engineering but also drastically reduces the reliance on extensive experimentation. Most importantly, When paired with this innovative tool, the predictive performance of GNNs has significantly improved.
That tool is getML.
Framework and Dataset
GetML is a high-performance machine learning framework to build regression and prediction models on any kind of relational data. While I briefly touch on getML’s prediction algorithms, I will mostly exploit its feature learner capabilities.
The dataset used for benchmarking is the well known CORA data set, that includes 2708 scientific publications classified into one of seven classes. The papers reference each other, giving rise to a graph network. Hence, in GNN terminology, the papers are nodes, the references are edges. Words in the papers’ abstract are features of the papers, i.e. node attributes.
Since its creation in 2000, the CORA dataset has facilitated the growth of an entire ecosystem of algorithms in NLP, RL and GNN. Papers with code documents and compares performance of cutting edge research on the dataset. As of 2023, the list is topped by an implementation attaining 90.16% in accuracy. This value is used as a benchmark.
Analysis
Since I am interested in simplicity and ease of implementation, I will start out with an off-the-shelf GNN implementation, compare its performance to the getML standalone solution, and then combine both approaches.
Standalone solutions
For full implementation details, please check out my Kaggle notebook. As a baseline model, I constructed a simple GNN embedded in the message passing framework. Using a graph convolutional operator (GCNConv) readily available in pytorch-geometric, I trained the GNN on 70% of the data and tested it on the remaining 30%. With this setting, 87.6% accuracy was attained. Next, I implemented a getML standalone solution. In this approach I capitalized on getML’s end-to-end prediction capabilities: getML’s “FastProp” feature learner and getML’s XGBoostClassifier to predict the labels. Randomizing over different train/test splits, this approach attains an accuracy of 88.3%.
Combine GNN and getML
Now, of course the next question arises: What if I combine the best of both worlds: The predictive power of Graph Neural Networks and FastProp’s ability to condense disparate data into valuable feature information?
So far, the word matrix one-hot-encodes 1433 words per paper’s abstract. That sparse representation does not only seem clumsy and inefficient, it is. getML’s FastProp algorithm aggregates it into a dense matrix, consisting of only 200 but highly discriminating features per node. Subsequently that 200 features were fed to the same off-the-shelf GNN as node attributes, just like before. Averaging over many train test splits, an astounding 92.5% accuracy was reached. Further results and programmatic implementation can be found in this notebook that investigates this symbiosis in full detail.
That means, without domain knowledge of linkage of academic papers with respect to their contents, without technical expertise of GNNs, and without any efforts and time investment in tinkering with parameters, any data science practitioner beats state of the art implementations by simply adding getML’s feature learner to the mix.
Consistency of Results
Now, one may object, perhaps I was simply lucky and picked by chance the right GNN configuration. I put this claim to the test and ran large scale trials probing different neural layer implementations. For every layer/optimisation combination I trained on different 30 randomized train test splits. Figure 1 shows how the resulting accuracies were distributed across conditions.
It is clear that for each and every layer implementation, the optimized features perform considerably better (each pairwise difference p < 0.001). The numerical values and the differences are shown in table 1. The improvement ranges from 4.1% points to 6.99% points, averaging 5.6% points.
What catches the eye is the remarkable consistency of results within the getML optimized condition. The standard deviation of the pooled results in the getML optimized condition with no preprocessing has 35% less spread than the condition without preprocessing. Considering the mere means within each layer condition (see table 1), spread in the getML condition shrinks by even 62%. That means, when using getML’s Feature Learner, it matters much less what layer implementation you end up choosing or how your data set is split, your result will be reliably good.
Conclusion
As a standalone solution, the getML pipeline slightly outperforms an off-the-shelf GNN implementation (88.3% vs 87.6%). Its true strength, however, lies in its flexibility to build on top of already high performing techniques. getML’s FastProp algorithm embeds a coarse one-hot-encoding in a dense matrix with highly distinguishing features. A simple GNN leverages this highly informative data in predictive accuracies well beyond the state of the art (92.5% vs 90.16%).
The information gain of FastProp’s automatic feature engineering is so significant, that predictive performance remains invariant across train test split randomization and variation of Neural Layer architectures. Computationally expensive experimentation, therefore is much less required.
And here I arrive at the gist of the matter: prepending getML’s feature learner to a Graph Neural Network leads to a
- massive improvement in accuracy, while
- eliminating extensive manual feature engineering and
- sharply reducing the need for computationally costly experimentation.
Whether you are seasoned machine learning engineer or a budding data scientist, getML’s feature learner will boost your models almost effortlessly.
Throughout the article, I used getML’s community edition. Being open source and free of charge it offers the already powerful, brute force based FastProp algorithm. The enterprise edition takes automated feature engineering then to the next level. Decision Tree based Feature Learning Algorithms (Multirel, Relboost, RelMT) produce more complex features for even higher predictive model performance.
To learn more about getML, check out its other use cases or give the community edition a try on your project. If building high-performance models is something you or your data science team cares about, contact me and my colleagues from getML at [email protected] to explore how we can help you.