What is it about?
In many real-world networks—like social networks, citation graphs, or biological systems—connected items (called "nodes") often have very different characteristics. This is known as heterophily. Traditional graph learning models, including popular Graph Neural Networks (GNNs) and graph Transformers, usually assume that connected nodes are similar (a property called homophily). This assumption doesn't hold in heterophilic scenarios, leading to poor performance. To solve this, we introduce FDphormer, a new kind of graph Transformer. Unlike existing methods, FDphormer doesn't just look at how nodes are connected—it also examines the differences between the features of connected nodes. We designed a new encoding method called DiSP, which captures these differences and helps the model better understand relationships in heterophilic graphs. We also explored the math behind how models generalize to new data and found that while adding this feature-difference information increases the model's power, it also risks overfitting. To address this, we add a regularization strategy to keep the model stable. Extensive experiments show that FDphormer performs very well on various types of graphs, especially those with heterophily, while remaining competitive on more traditional (homophilic) data.
Featured Image
Photo by Logan Voss on Unsplash
Why is it important?
In recent years, graph Transformers have become popular for understanding complex network data, such as social networks, molecules, or knowledge graphs. However, most of these models are still built around the assumption that connected nodes are similar. This works well for some datasets, but breaks down in real-world cases where connected nodes are actually very different—a common situation known as heterophily. What makes our work unique is that we directly address this challenge. We propose a new position encoding method called DiSP, which measures not just the structure of the graph, but the feature differences between connected nodes. This allows the model to better distinguish between meaningful and misleading relationships—something previous methods largely ignored. Our work is also timely, as interest in applying Transformers to graph-structured data is growing rapidly. But many current models still struggle in heterophilic settings like protein interaction networks, social media, or recommendation systems. FDphormer bridges this gap, providing a simple and efficient way to boost performance where it’s needed most. By combining theoretical analysis with practical improvements, our model is not just a new tool—it could reshape how graph learning models are designed for diverse and challenging datasets.
Perspectives
From my perspective, this work represents a meaningful step forward in addressing a long-standing limitation in graph learning. As someone who's spent time exploring the boundaries of GNNs and Transformers, I’ve often noticed how models underperform in heterophilic graphs, which ironically are more common in the real world than we tend to admit. What excited me most during this project was the realization that something as "simple" as capturing the difference between node features—rather than just their connections—could make such a big impact. The idea of encoding this feature difference into the model's attention mechanism felt intuitive yet surprisingly overlooked in the literature. Working on FDphormer was also a chance to blend theory and practice: developing a model that performs well empirically, while also offering solid theoretical guarantees. Personally, I see this as the kind of research that can influence future graph model design—not just by improving accuracy, but by shifting how we think about structure and information in complex networks.
Dong Li
Harbin Institute of Technology
Read the Original
This page is a summary of: FDphormer: Beyond Homophily with Feature-Difference Position Encoding, ACM Transactions on Knowledge Discovery from Data, April 2025, ACM (Association for Computing Machinery),
DOI: 10.1145/3727882.
You can read the full text:
Contributors
The following have contributed to this page







