2025 年 25 巻 p. 36-52
Accurate prediction of drug response in cancer treatment remains a critical challenge due to the complex biological interactions underlying tumor sensitivity and resistance. In this work, we introduce OT-GNN, a novel graph neural network framework that leverages optimal transport theory to integrate prior drug-target interaction knowledge with gene expression profiles for interpretable and robust drug response prediction. By embedding an optimal transport-based alignment mechanism into the GNN architecture, OT-GNN dynamically reweights gene importance tailored to each drug–cell line pair, enhancing both predictive accuracy and biological interpretability. We evaluate OT-GNN on a processed NCI-60 dataset under zero-shot learning settings, demonstrating superior performance compared to traditional machine learning models, recent deep learning methods, and standard GNN variants without our proposed alignment. OT-GNN achieves state-of-the-art ROC-AUC and PR-AUC scores, with improved stability across multiple runs, highlighting its potential as a reliable tool for precision oncology applications. Our approach bridges the gap between data-driven modeling and biological prior knowledge, providing a pathway toward more transparent and effective drug response prediction.
1. Introduction
Precision oncology aspires to tailor cancer treatments to the molecular characteristics of individual patients, relying heavily on the ability to predict how tumors will respond to specific drugs. The proliferation of high-throughput pharmacogenomic resources, including the Cancer Cell Line Encyclopedia (CCLE) [1], the Genomics of Drug Sensitivity in Cancer (GDSC) [2], and the NCI-60 panel [3], has enabled systematic investigations of the relationship between genomic profiles and drug response. These datasets provide a fertile ground for machine learning (ML) approaches to model complex genotype-phenotype interactions [4]. However, drug response prediction remains a fundamentally difficult problem due to the inherent heterogeneity of tumor biology, sparsity of labeled data, high dimensionality of input features, and variability across cell lines and compounds [5].
Early efforts in drug response modeling relied on classical machine learning algorithms, such as support vector machines, random forests, and gradient boosting [6, 7]. While these approaches offer robust performance in controlled scenarios, they often treat genes as independent features, failing to model the complex interactions and regulatory dependencies between them. Moreover, these models lack biological interpretability, a critical limitation in high-stakes biomedical applications. To overcome these challenges, recent advances have embraced deep learning techniques that incorporate biological priors. In particular, graph neural networks (GNNs) have shown promise in integrating multi-omics data with prior knowledge in the form of biological networks, such as protein-protein interaction (PPI) graphs [8, 9]. These models capture the non-Euclidean structure of biological systems and exploit message passing mechanisms to learn expressive representations of gene interactions. Nevertheless, existing GNN-based models often rely on static graph structures and assume uniform importance across all nodes and edges, disregarding sample-specific biological signals. Additionally, their predictions are typically opaque, offering limited insight into which genes or pathways drive a given response.
To address these limitations, we propose OT-GNN, a novel graph-based framework that integrates optimal transport theory into GNN architecture for interpretable and effective drug response prediction. Our key innovation lies in leveraging optimal transport (OT) to align cell line-specific gene activity profiles with drug-target interaction priors [10,11]. This alignment produces an optimal coupling matrix that reflects the biological relevance of each gene in the context of a given drug and cell line. By embedding this transport plan into the GNN's message-passing mechanism, OT-GNN dynamically reweights gene contributions based on their biological importance, enabling both enhanced prediction and interpretability.
Unlike many existing models that apply a fixed graph structure or depend heavily on edge-level features, OT-GNN constructs a sample-specific attention mechanism rooted in biological priors, without introducing additional graph complexity. This leads to a more flexible and biologically grounded model that not only adapts to each input sample but also remains stable and computationally tractable. Importantly, the learned transport plans offer a principled way to interpret the model's predictions, as they highlight the genes most responsible for sensitivity or resistance to a particular drug.
We validate our model using a processed version of the NCI-60 dataset [12], applying zero-shot learning settings to test generalizability across unseen drugs and cell lines. Our experimental results demonstrate that OT-GNN outperforms traditional machine learning models, prior deep learning baselines, and GNN variants without the OT-based alignment mechanism. The improvements are particularly pronounced in terms of ROC-AUC and PR-AUC—two key metrics in imbalanced classification tasks [13]. Moreover, OT-GNN shows strong performance stability across multiple runs and data splits, reinforcing its robustness and reliability.
Our main contributions are as follows:
By grounding predictions in biological knowledge and maintaining interpretability, OT-GNN moves beyond black-box deep learning approaches and offers a step toward more trustworthy, generalizable, and clinically actionable models in computational oncology.
2. Methods
2.1 Optimal Transport Foundations
Optimal transport (OT) [14,15,16,17,18] is a mathematical framework concerned with determining the most cost-efficient way to move mass—such as probability distributions—from one configuration to another. The goal is to minimize the total transportation cost between two distributions.
Let us define two discrete point sets, 𝒳
The transportation plan is represented by a non-negative matrix
| (1) |
| (2) |
Given a cost matrix
| (3) |
where the admissible set of transport plans is
Here,
When
| (4) |
We solve this linear program using the simplex method [19], iteratively updating feasible solutions until optimality criteria are met.
2.2 Gene-Gene Interaction Graph Construction
To build a biologically grounded gene-gene interaction network enriched with informative node features and initial importance scores, we integrated data from three sources:
Gene expression values were normalized via Transcripts Per Million (TPM), then
We define the initial Drug-Target Interaction (DTI)-based importance score between drug
Here,
Here,
2.3 Drug-target and Gene Activity Signals
To complement the drug-target signal, we also constructed a gene activity profile representing the functional state of genes in each cell line. For each gene 𝒈𝒋 , we combined four types of molecular measurements: gene expression (
| (5) |
Here, the weights
2.4 Profile Matching via Optimal Transport
Gien a drug 𝒅 and a corresponding prior importance vector
We normalized both vectors to define discrete probability distributions:
| (6) |
Then, we computed the OT distance (e.g., Sinkhorn divergence) between
| (7) |
where
This OT distance 𝒲 𝜖(𝝻, 𝞶) serves as a biologically motivated compatibility score: a smaller distance suggests that the drug’s target genes are functionally active in the cell line, making the drug more likely to elicit a strong response. 𝒲 𝜖(𝝻, 𝞶) is a new vector over genes that represents how the drug’s target profile aligns with the gene activity profile of the cell line.
From the optimal transport plan 𝑻*, we derive the contextual alignment vector 𝑨𝑶𝑻 ∈ 𝑹|𝑮| by aggregating mass assigned to each gene across all drug targets:
Thus, 𝒲 𝜖 is a scalar value used in the loss function to measure the alignment cost, whereas 𝑨𝑶𝑻 is a gene-level importance vector used within the model to provide contextual biological signals to the graph neural network. These two quantities are related through the optimal transport plan 𝑻*, but serve distinct roles in the framework.
We incorporated this OT-guided score into the GNN training pipeline in two ways: (1) as an additional input feature, appended to the learned graph-level embedding; and/or (2) as a regularization term encouraging node importance propagation to be consistent with OT alignment. This integration enables the model to learn not only from structural relationships in the gene network but also from distributional alignment between drug action and cell context.
2.5 OT-Guided GNN Architecture
Our proposed model predicts drug response and uncovers gene-level importance by integrating prior drug-target knowledge with cell-specific molecular profiles using a graph neural network enhanced by optimal transport. We operate on a gene interaction network 𝑮=(𝑽, 𝑬) with |𝑽| genes and use node features
Given a drug 𝒅, the model receives as input: (i) a prior importance score vector
2.5.1 Contextual Attention Refinement Layer (CAR Layer)
The Contextual Attention Refinement (CAR) Layer is designed to jointly refine node feature representations and node-level importance scores by integrating local graph topology with contextual drug-cell line alignment signals. Its job is to contextualize the importance signal with respect to the local network topology and gene activity. The output consists of an updated feature matrix
We enhance node features with contextual alignment information by concatenating the OT-derived alignment vector
| $$\boldsymbol{X}^{\boldsymbol{'}} = [\boldsymbol{X} | \boldsymbol{A}_{\boldsymbol{OT}}]$$ |
We then apply a graph neural network layer, such as GAT or TransformerConv, to propagate contextualized features over the graph:
where
To balance structural and contextual relevance, we introduce two types of attention:
These are computed as:
where
Node features are updated by aggregating messages from their neighbors, weighted by structural attention and modulated through a contextual gating mechanism:
This step yields the updated feature matrix
Refined node-level importance scores are computed from the updated features:
This produces the updated alignment vector
To enhance interpretability, a smooth thresholding function transforms the raw importance:
The layer returns both

Figure 1 Overview of the OT-GNN framework for drug response prediction
The pipeline begins by constructing an attributed gene–gene interaction graph from the PathwayCommons database, where each node is enriched with multi-omics-derived gene activity features. In parallel, drug–cell line contextual alignment is computed via Optimal Transport (OT) between drug-target profiles and cell-specific gene activity distributions, yielding an importance vector that highlights contextually relevant genes. Both the attributed graph and the OT-derived importance scores are input into the CAR-GNN model, which features a Contextual Attention Refinement (CAR) layer. This layer integrates structural and biological signals using a dual attention mechanism, refines node features and importance scores, and propagates them through stacked GNN layers. A final graph-level embedding is aggregated and passed through a prediction head to estimate drug response. The model is trained using a composite loss that balances prediction accuracy, importance sparsity, and OT-based biological alignment.
2.5.2 Model Architecture and Prediction Head
The model stacks three IP Layers, each followed by GraphNorm, ReLU, and Dropout. The graph-level representation is obtained by global mean pooling of the final node embeddings:
To predict drug response, we use a fully connected layer followed by a sigmoid activation:
2.5.3 Loss Function
The training objective is a composite loss that balances predictive accuracy, interpretability, and biological plausibility:
| $$\mathcal{L} = \underbrace{\mathcal{L}_{\text{BCE}}(\hat{y}, y)}_{\text{Prediction loss}} + \underbrace{\lambda_{\text{imp}} \cdot \|\mathbf{I}^{\text{final}}\|_{1}}_{\text{Importance sparsity}} + \underbrace{\lambda_{\text{OT}} \cdot \mathcal{W}_{\epsilon}(\mu, \nu)}_{\text{OT alignment regularization}}$$ |
where:
This joint objective encourages accurate prediction, interpretable sparsity in importance scores, and biologically coherent alignment between drug targeting and cell context.
3. Experiments
3.1 Datasets
We based our experiments on the NCI-60 pharmacogenomic dataset [3], retrieving multi-omics profiles and drug sensitivity data using the rcellminer package [27]. Drug response was quantified using log-transformed GI50 values, which we binarized using a threshold of -4.6 to distinguish sensitive from resistant responses, yielding a near-balanced label distribution. After filtering for compounds annotated with NSC (National Service Center) identifiers, our final dataset included 52,000 drug-cell line interaction pairs, of which 34,000 were labeled as sensitive and 18,000 as resistant.
To evaluate the model under a zero-shot prediction setting, we designed a split that excluded 30% of cell lines and 40% of drugs from training. Specifically, 560 drugs and 40 cell lines were used for training and validation, while 370 drugs and 20 cell lines were held out for testing. This partitioning produced approximately 17,500 training instances, 4,200 for validation, and 6,300 for testing.
3.2 Baseline Methods and Hyperparameter Optimization
To comprehensively assess the performance of our proposed OT-GNN framework, we benchmark against a diverse suite of baseline models, spanning traditional machine learning methods, deep learning architectures, and graph neural networks. All models were evaluated using the same input features (multi-omics and DTI-based gene features) and data splits for a fair comparison. We applied systematic hyperparameter tuning for all models using the Optuna framework [28], optimizing validation performance through randomized or grid search strategies.
All models were trained using the Adam optimizer with early stopping on the validation set. Performance was averaged over five independent runs using fixed train/validation/test splits. We report mean and standard deviation for ROC-AUC, PR-AUC, accuracy, precision, and specificity. Input features were consistently normalized across models and included expression, mutation, methylation, CNV, and initial DTI-based signals. This unified evaluation pipeline ensures fair and robust comparison.
In our framework, the base GNN used within the Contextual Attention Refinement (CAR) layer can be flexibly instantiated with different architectures to examine their effect on performance. OT-GNN (GAT) uses the Graph Attention Network (GAT), OT-GNN (GT) employs a Graph Transformer architecture (GT), OT-GNN (GINE) integrates the Graph Isomorphism Network with Edge features (GINE). In all cases, the CAR layer applies these GNNs to the contextualized feature matrix and edge index to generate updated node embeddings and propagate importance scores. This modularity allows us to compare architectural performance under a unified interpretable optimal transport framework.
3.3 Evaluation Metrics
To comprehensively assess the predictive performance of our model and baselines, we report a set of widely used evaluation metrics for binary classification tasks:
In Table 1, we use upward arrows (↑) to indicate that higher values for these metrics are desirable. For each metric, the best-performing method is highlighted in bold.
3.4 Prediction Performance
We assessed the performance of our proposed model, OT-GNN, against a variety of benchmarks, including five classical machine learning models, two state-of-the-art baselines, and three graph neural network (GNN) variants without our proposed optimal transport-guided importance propagation. Table 1 summarizes the results, reporting the mean and standard deviation over five independent runs.
OT-GNN demonstrates consistent and superior performance across nearly all evaluation metrics, especially on ROC-AUC and PR-AUC, which are critical in imbalanced classification scenarios. Among all variants, OT-GNN (GT) achieved the best ROC-AUC (0.801) and PR-AUC (0.901), indicating its strong discriminative power and high precision under class imbalance.
Compared to its non-IP counterpart, OT-GNN (GT) shows a 3.8% improvement in ROC-AUC and a 4.3% improvement in PR-AUC, highlighting the substantial contribution of optimal transport-guided importance propagation. Likewise, the GINE-based variant saw gains of 5.4% in PR-AUC and 5.4% in ROC-AUC, further validating the module’s adaptability across architectures.
While DeepDSC slightly outperforms in raw accuracy and precision, our model offers a more balanced performance profile, with considerably better PR-AUC and ROC-AUC. This reflects OT-GNN’s ability to generalize effectively rather than overfitting to the dominant class.
Interestingly, MOFGCN, while exhibiting the highest specificity (0.893), suffers from poor overall classification ability (ROC-AUC: 0.487), confirming its bias toward predicting resistance only. In contrast, OT-GNN provides a more even predictive capability.
Finally, all variants of OT-GNN exhibit low standard deviations across runs, suggesting that the model is robust and stable, even under variations in data split and initialization—an important characteristic for real-world biomedical applications.
Table 1. Predictive Performance Comparison for Binary Classification
Results are averages over 5 runs with standard deviations. Best values per metric are in bold.
3.5 Interpretability Analysis
Our model enhances interpretability by assigning each gene a context-dependent importance score, which reflects its influence on the predicted drug response. These scores are derived from the Contextual Attention Refinement (CAR) layers, which propagate drug-target information across the gene-gene network while integrating cell line–specific activity via optimal transport (OT) alignment. This design enables the model to highlight both direct and indirect contributors to drug efficacy in a biologically informed manner. To evaluate the interpretability of the proposed OT-GNN framework, we investigate how the model assigns gene-level importance scores during drug response prediction. We present two representative case studies illustrating the biological relevance of the model's outputs.
3.5.1 Case Study I: Bortezomib
Bortezomib is a proteasome inhibitor used in multiple myeloma and certain lymphomas. Table 1 lists the top 10 genes identified for Bortezomib, a proteasome inhibitor used in multiple myeloma. The top-ranked gene, PSMB5, a catalytic subunit of the 20S proteasome, is the known molecular target of Bortezomib, validating the model's alignment with established pharmacological knowledge. Genes such as RELA and NFKBIA are involved in the NF-κB pathway, which is regulated by proteasomal degradation and is central to Bortezomib’s mechanism of inducing apoptosis. Other genes like HSPA1B and RPL10, though not direct targets, are related to protein folding and stress response pathways, possibly influencing cellular susceptibility to proteasome inhibition. The identification of these genes highlights OT-GNN’s ability to capture secondary mechanisms that influence treatment response, offering interpretability beyond known direct targets.
Table 2. Top 10 important genes for Bortezomib and associated evidence
3.5.2 Case Study II: Erlotinib
We further evaluated the model on Erlotinib, an EGFR tyrosine kinase inhibitor that targets EGFR, used to treat non-small cell lung cancer. As shown in Table 2, the top-ranked gene, EGFR, the known direct molecular target of Erlotinib, demonstrates strong attribution fidelity and the model’s ability to recover established drug-target interactions. Additional genes such as PIK3CA, GRB2, and STAT3 are part of the EGFR downstream signaling network and are known to affect Erlotinib sensitivity or resistance. Notably, TP53 and CDKN2A are frequently mutated in cancer and modulate cell cycle response and apoptotic pathways that intersect with EGFR signaling. The ability of the model to recover both direct and contextually relevant genes highlights its interpretability and biological validity.
Table 3. Top 10 important genes for Bortezomib and associated evidence
The rankings in Tables 2 and 3 are based on the normalized importance scores produced by the CAR (Contextual Attention Refinement) layer. Specifically, after the graph message passing and dual attention steps, each gene node is assigned a scalar importance score
These case studies highlight OT-GNN’s capacity for interpretable prediction by assigning biologically meaningful importance scores to genes, both known targets and secondary effectors. The consistency of top-ranked genes with known drug mechanisms and resistance pathways demonstrates that the model integrates prior knowledge with contextual cellular signals effectively. This interpretability may aid in identifying predictive biomarkers, suggesting alternative pathways for therapeutic targeting, and generating mechanistic hypotheses for experimental validation.
4. Conclusion
In this study, we presented OT-GNN, a novel graph neural network framework that incorporates optimal transport to effectively integrate prior drug-target interaction data with heterogeneous multi-omics profiles for interpretable drug response prediction. By aligning drug-target importance signals with cell line-specific gene activity distributions, our method dynamically adjusts gene importance within the GNN, enhancing both predictive performance and biological interpretability. Extensive experiments on the NCI-60 dataset demonstrate that OT-GNN consistently outperforms traditional machine learning models, existing deep learning approaches, and baseline GNN architectures lacking optimal transport-guided alignment. Furthermore, the stability of our model across multiple runs underscores its robustness in handling complex and imbalanced pharmacogenomic data. Future work may extend this framework to incorporate additional omics modalities and explore its applicability to other precision medicine challenges, further advancing the integration of prior biological knowledge with data-driven modeling for improved clinical decision support.
5. Code Availability
To support transparency and reproducibility, we plan to release the source code for OT-GNN upon publication. The repository will include implementation details, training scripts, and instructions for reproducing the main experiments. The code will be made publicly available at: https://github.com/maburidi/OT-GNN