Predict lncRNA-drug associations based on graph neural network

1 Introduction

LncRNAs are a class of non-coding RNAs transcribed from DNA with a length of over 200 nucleotides (Ponting et al., 2009). They account for 70%–80% of all non-coding RNAs and play crucial regulatory roles in numerous cellular processes, including but not limited to transcription, splicing, translation, DNA repair, and regulation of genes. The quantity and biological importance of lncRNAs determine their widespread involvement in all physiological activities of living cells and the pathogenesis of human diseases, such as cancer, Parkinson’s disease, and cardiovascular disease (Riva et al., 2016; Schmitz et al., 2016; McCabe and Rasmussen, 2021). LncRNAs represent a new type of potential therapeutic targets, that can affect the diagnosis, treatment, and prognosis of diseases, and have attracted significant attention (Blokhin et al., 2018; Fernandes et al., 2019; Winkle et al., 2021).

Due to the key roles of lncRNAs in diseases, it is crucial to develop lncRNA-targeted drugs and technologies. This presents a significant opportunity for the treatment of lncRNA-related diseases and represents a new area for drug development (Sangeeth et al., 2022). Emerging studies have shown that small-molecular drugs can inhibit the proliferation of tumor cells or tumor stem cells by regulating the expression of lncRNAs, laying a crucial theoretical foundation for the advancement of lncRNA-targeted therapeutics (Liu et al., 2021). To develop lncRNA-targeted drugs, it is necessary to take three preparatory steps: elucidating the action mechanism of lncRNAs in diseases, analyzing their structural and functional pockets, and finding small-molecular drugs that can bind specifically to the pockets. One important aspect of this process is identifying associations between lncRNAs and drugs (Jiang et al., 2019; Chen Y. et al., 2021). Predicting lncRNA-drug associations (LDAs) not only facilitates the selection of potential drug candidates but also streamlines the drug discovery process, ultimately propelling the realization of efficacious lncRNA-targeted therapies and fostering advancements in precision medicine.

LDAs are mainly identified through biological experiments. For example, curcumin plays a crucial role in the treatment of various cancers by regulating lncRNAs (Patel et al., 2020). It inhibits the expression of lncRNA H19, and restores MEG3 levels via demethylation, thus enhancing the sensitivity of cancer cells to chemotherapeutic drugs (Zhang et al., 2017; Cai et al., 2021). Wang et al. (2020) confirmed that the oncogenic factor lncRNA CCAT2 was overexpressed in ovarian cancer, and calcitriol, the vitamin metabolite, can inhibit the proliferation, migration, and differentiation of ovarian cancer cells by inhibiting the expression of CCAT2. However, identifying LDAs based on biological experiments is time-consuming and costly, there is a need for efficient and accurate computational methods to predict potential LDAs, which can be further verified by biological experiments.

Jiang et al. (2019) identified LDAs based on the hypothesis that lncRNAs with similar sequences are often regulated by the same drug, and drugs with similar structures tend to regulate the same lncRNA. Wang et al. (2018) utilized Elastic Network (EN) regression to predict potential LDAs by integrating lncRNA expression profiles and drug response data in cancer cells. However, the limitation of these methods is that they heavily rely on specific features of the existing data, which may affect the prediction performance. Although predicting LDAs is receiving increasing attention, the relevant prediction methods are still relatively lacking.

At present, a wealth of computational methods has been accumulated for predicting small molecule drug-miRNA associations (Qu et al., 2019; Yin et al., 2019; Chen et al., 2020). Considering that both lncRNAs and miRNAs are non-coding RNAs involved in gene expression regulation and cellular functions, and share similar regulatory mechanisms (Yan and Bu, 2021), the methods for predicting drug-miRNA interactions hold significant implications for predicting LDAs. Zhao et al. (2020) developed a model using symmetric nonnegative matrix factorization and Kronecker regularized least squares to predict small molecule drug-miRNA associations. Chen et al. (2021) predicted small molecule drug-miRNA associations based on bounded nuclear norm regularization. Wang et al. (2022) presented an ensemble of kernel ridge regression-based method to identify potential small molecule-miRNA associations. Niu et al. (2023) employed a combination of GNNs and Convolutional neural networks (CNNs) to predicted small molecule drug-miRNA association.

Recently, due to the breakthroughs in deep learning and the huge improvements in computing power, models based on deep learning, particularly those employing GNNs, have been applied in multiple bioinformatics-related tasks, such as lncRNA-disease association prediction, and drug-target interaction prediction (Xuan et al., 2019; Chen et al., 2020; Kumar Shukla et al., 2020; Zhao et al., 2023). Yin et al. (2023) proposed a general framework using residual GCN and CNNs to predict drug-target interactions. Wang and Zhong (2022) proposed a method (gGATLDA) to identify potential lncRNA-disease associations based on graph attention networks (GAT). The success of the above GNNs-based methods can be attributed to the three main reasons: 1) biological correlations can be modeled naturally as graph structures, 2) GNNs have the advantage of capturing complex network relationships, 3) the introduction of attention mechanisms enables the model to focus locally on important nodes in the graph and effectively integrate node information on a global scale.

Therefore, based on the experiments validated LDAs dataset (D-lnc (Jiang et al., 2019)), we propose a GNNs-based framework to predict LDAs by referring to the gGATLDA method which was originally designed to predict lncRNA-disease associations. In this paper, we first extract the lncRNA-drug bipartite graph according to the LDAs matrix and obtain one-hop enclosing subgraphs of all lncRNA-drug pairs from the bipartite graph. Then, the feature vectors of lncRNA-pairs are constructed according to Gaussian interaction profile kernel lncRNA (drug) similarities. Finally, GCN learns lncRNA and drug node embeddings and obtain local spatial characteristics of nodes. GAT uses the attention mechanism to integrate the global information of the lncRNA-drug bipartite graph. Our model takes full advantage of GCN and GAT to predict novel LDAs. Results show that the method achieves high AUC and AUPR, and the ablation experiments show that our model performs better than GCN and GAT. The case studies on two drugs (Berberine and Panobinostat) and two lncRNAs (NEAT1 and MEG3) demonstrate the effectiveness of the model in predicting the potential LDAs. In the functional enrichment analysis, we further verified the validity of our predicted LDAs from the perspective of the relationship between the biological function of drugs and the enrichment pathway of lncRNA target genes. All these results suggest that the framework used in this study is an efficient method for predicting LDAs.

2 Materials and methods2.1 Materials

Three benchmark datasets including the Gene Expression Omnibus (GEO) dataset, Connectivity Map (cMap) dataset, and Validated dataset (Jiang et al., 2019) are downloaded from http://www.jianglab.cn/D-lnc/index.jsp. The cMap dataset and GEO dataset were obtained by re-annotating microarray probes in cMap and GEO databases, respectively, and screening lncRNA differential expression data before and after drug therapy. The Validated dataset was created by searching experimentally verified drug modification of lncRNA expressions. We obtain three benchmark datasets (Dataset 1, Dataset 2, and Dataset 3) by removing the repeated LDAs of GEO dataset, cMap dataset, and Validated dataset respectively. As the number of lncRANs and drugs in three benchmark datasets is unbalanced, only when Dataset 1 and Dataset 2 are combined, the number of them is relatively balanced. Therefore, we combine Dataset 1 and Dataset 2 into Dataset 4, which is used as a training dataset in the case study to predict LDAs in Dataset 3 (see case study section). Dataset 5 is merged from Dataset 1, Dataset 2, and Dataset 3. Table 1 shows the detailed information of five datasets. We treat the known LDAs as positive samples and randomly select the negative samples with the same number of positive samples from the unknown LDAs.

www.frontiersin.org

Table 1. The detailed information of five datasets.

2.2 Methods

The flowchart of the method is shown in Figure 1. Firstly, construct the lncRNA-drug association matrix, lncRNA similarity matrix, and drug similarity matrix. Secondly, construct one-hop enclosing subgraphs according to the lncRNA-drug bipartite graph, and obtain lncRNA node features and drug node features, respectively. Further, the one-hop enclosing subgraph of each lncRNA-drug pair and their feature vectors are input to the GNNs model. Finally, the lncRNA vector and drug vector are concatenated and processed by Softmax to obtain the prediction score.

www.frontiersin.org

Figure 1. The flowchart of our method. (A) The LDAs matrix, lncRNA similarity matrix, and drug similarity matrix are constructed, respectively. (B) Obtain lncRNA-drug bipartite graph and one-hop subgraphs, and construct the initial feature vector of lncRNA and drug. (C) Extract feature representation of lncRNAs and drugs based on GCN and GAT. (D) Concatenate the lncRNA vector and drug vector to obtain a new vector, and the prediction score is obtained by Softmax.

2.2.1 Constructing similarity matrices for lncRNAs and drugs

Because of the sparsity of the lncRNA-drug association matrix LD∈Rm×n, we calculate lncRNA similarity LSli,lj and drug similarity DSdi,dj by the following Gaussian interaction profile kernel (GIP) (van Laarhoven et al., 2011; Yang and Li, 2021):

LSli,lj=exp−rlIPli − IPlj2(1)DSdi,dj=exp−rdIPdi − IPdj2(2)

where LS∈Rm×m and DS∈Rn×n denote the lncRNA similarity matrix and drug similarity matrix, respectively. m and n represent the number of lncRNAs and drugs respectively. IPli and IPdi are binary vectors, which represent the i th row and i th column of LD, respectively. If lncRNA li is associated with drug dj, LDi,j=1, otherwise LDi,j=0. rl and rd are used to adjust the kernel bandwidth, which are calculated as followed:

rl=1/1m∑i=1mIPli‖2(3)2.2.2 Extracting one-hop enclosing subgraph

The bipartite graph G is constructed from the matrix LD, where each known lncRNA-drug pair corresponds to an edge connecting the lncRNA li and drug dj, for unknown LDAs, there are no edges between li and dj. The one-hop enclosing subgraph G1V1,E1 of each lncRNA-drug pair li,dj can be defined as following: V1 is the set of nodes containing one-hop neighbor nodes of li, one-hop neighbor nodes of dj, as well as node li and node dj, and E1 is edge set. Each node in subgraphs can be labeled to distinguish its role (Zhang and Chen, 2020). We use 0 and 1 to label target lncRNA node and target drug node, respectively, and use 2i and 2i+1 to label the one-hop neighbor nodes of li and the one-hop neighbor nodes of dj, respectively, where i represents the order of neighbor nodes, and it is set to 1 according to gGATLDA (Wang and Zhong, 2022).

2.2.3 Constructing and denoising original feature vectors

The original lncRNA and drug feature vectors are constructed from lncRNA similarity matrix and drug similarity matrix, respectively. However, due to the high dimension of original features and the sparsity of lncRNA (drug) similarity matrix, we employ principal component analysis (PCA) for dimension reduction. PCA is a classical, efficient, and unsupervised feature selection method, which can not only retain as much feature information as possible while reducing the feature dimension but also greatly reduce the training time of the model. Assume the original lncRNA and drug feature vector fl0=fl10,fl20,fl30,...,flm0 and fd0=fd10,fd20,fd30,...,fdn0, respectively. After performing PCA, we obtain the feature vectors fl′=fl1′,fl2′,fl3′,...,fla′ and fd′=fd1′,fd2′,fd3′,...,fdb′, where a and b denote the feature vector dimension of lncRNAs and drugs, respectively. Since a and b may not be equal, in order to ensure that the input feature dimensions of each node are the same, we take the sum of 4+a+b as the feature dimensions of the nodes, where the first 4 dimensions represent the one-hot encoding of the node labels to distinguish the roles of different nodes. The extra b dimensions of lncRNA node features and a dimensions of drug node feature are filled with 0 values. We construct the lncRNA feature matrix Flnc=Rm×4+a+b and the drug feature matrix Fdrug=Rm×4+a+b. The feature vector of lncRNA l is fl=p1,p2,p3,p4,fl1′,fl2′,fl3′,...,fla′,0,0,0,...,0, and pj1≤j≤4 represents the one-hot encoding of the node label, distinguishing different roles. Similarly, the feature vector of a drug d is fd=p1,p2,p3,p4,0,0,0,...,0,fd1′,fd2′,fd3′,...,fdb′.

2.2.4 The model based on GCN and GAT

GNNs are a class of data models for processing graph structures that utilize a message-passing mechanism to update the node embeddings. GCN and GAT are two specific GNN models whose core idea is to update node embeddings by aggregating neighbor nodes’ information. The difference is that GAT introduces an attention mechanism during the message-passing process, which can adaptively assign different weights to different nodes, allowing for more flexibility in capturing relationships between nodes.

In the model, the GCN layer is employed to update the features of lncRNA and drug nodes. The feature representation of node i in the k+1 layer is presented as follows:

hik+1=σ∑j∈Ni∪i1di·djhjkWk+1(5)

where hjk represents the feature vector of node j in layer kk=0,1,2,3,...,n, dj denotes the degree of node j, Ni represents the set containing all neighbors of node i, and Wk+1 denotes the parameter matrix to be learned in the k+1 GCN layer.

To get the weights between different nodes, GAT introduces the attention coefficient. The attention coefficient between node i and node j is calculated as follows:

where a represents a shared attention mechanism to calculate the attention coefficient, and σ represents the LeakyReLU activation function.

To compare the attention coefficient between different nodes, the normalized attention coefficient αij is calculated as follows:

αij=softmaxeij=expeij∑m∈Niexpeim(7)

After obtaining the attention coefficients between node i and its neighbor nodes, we obtain the final representation hik+1 by taking a weighted summation of its neighbor nodes, and it is calculated as follows:

hik+1=σ∑j∈NiαijWkhjk+1(8)

where σ represents ELU activation function.

2.2.5 Prediction score

For the final output of the last GAT layer, the vector representations of target lncRNA and target drug are concatenated:

where hli and hdj denote the final feature representation of the lncRNA li and dj, respectively. The purpose of concatenating the feature vectors of target lncRNA and target drug is to integrate the feature information of lncRNA and drug node pairs to form a richer feature representation and to reduce information loss to a certain extent. This can help the model better understand the relationship between lncRNAs and drugs and improve the accuracy of prediction.

Finally, for the representation fli,dj, we use Softmax as an activation function to obtain the prediction score y′li,dj:

yli,dj′=Softmaxfli,dj=efli,dj∑j=1nefli,dj(10)

The binary cross-entropy loss function is used to train the weight Wk:

Loss=−yli,dj⁡logy′li,dj+1 − yli,djlog1 − y′li,dj(11)

where yli,dj represents the real label.

2.3 Evaluation criteria

In this study, we evaluate the performance of the model by means of AUC (Area Under Curve), AUPR (Area Under the Precision-Recall curve), precision, accuracy, F1-Score, and recall. AUC means the area under the Receiver Operating Characteristic (ROC) curve, which is plotted by the true positive rate (TPR) and the false positive rate (FPR) at different thresholds. The TPR and FPR are calculated as follows:

where TP and TN are the numbers of correctly identified positive and negative samples respectively. FP and FN are the numbers of misidentified positive and negative samples, respectively.

In addition, the evaluation metrics including precision, recall, F1-score, and accuracy are calculated as follows:

F1−score=2×precision×recallprecision+recall(16)Accuracy=TP+TNTP+TN+FP+FN(17)
3 Results

In this section, firstly, we select the appropriate parameters of the model through parameter optimization. Secondly, we show the experimental results of six evaluation metrics on five datasets and conduct ablation experiments on five datasets. Thirdly, case studies are conducted on two drugs and two lncRNAs, which aim to validate the ability of the method to predict potential LDAs. Finally, we performed the KEGG functional analysis based on the results of case studies to further verify the validity of the predicted LDAs, especially for those that are unconfirmed in the case study.

3.1 Parameter optimization

We initially explore the influence of various hyperparameter combinations on the performance of predicting LDAs across five datasets. These hyperparameters include epochs, batch size, learning rate, the initial feature vector dimensionality selected by PCA, and the layers of GCN and GAT, respectively. We utilize grid search to tune these six hyperparameters. The epochs range extends from 10 to 50, incremented by 10. Batch size is selected from the set , learning rate is chosen from , and the initial feature vector dimensionality selected by PCA ranges from . The experimental results are shown in Figure 2, the optimal model performance is attained when a particular combination of hyperparameters is used on Dataset 1 to Dataset 5. Specifically, epochs are set to , batch sizes to , and learning rate uniformly to 0.001. Additionally, the number of initial PCA-selected feature dimensions is set to for each dataset.

www.frontiersin.org

Figure 2. Hyperparameter optimization results for five datasets. (A) The AUC values under the different epochs on five datasets. (B) The AUC values under the different batch sizes on five datasets. (C) The AUC values under the different learning rates on five datasets. (D) The AUC values under the different dimensions selected by PCA on five datasets.

In addition, the selection of model structure is crucial for prediction performance. Therefore, we also investigate the impact of the two main hyperparameters, the number of GCN and GAT layers on the model in Dataset 1, Dataset 2, and Dataset 3. Specifically, we select layer numbers ranging from 1 to 3 for both GCN and GAT. The increase in the number of GNN layers means aggregating features from higher-order neighbor nodes, but it may lead to the loss of local structure, resulting in overfitting and decreased prediction performance. For instance, as shown in Table 2, in Dataset 1, increasing the GCN layer from 1 to 2 improves performance, but further increasing it to 3 leads to a decrease in performance. It is worth noting that the model achieves the best performance across all three datasets when the model structure consists of 1 GCN layer and 3 GAT layers. Therefore, we choose 1 GCN layer and 3 GAT layers as our model architecture.

www.frontiersin.org

Table 2. The performance using different numbers of GCN and GAT layers on Dataset 1, Dataset 2, and Dataset 3.

3.2 Prediction performance of the model

To avoid potential random bias, we repeat the five-cross-validation process 100 times and average the performance metrics to derive the final results. The results are shown in Table 3, the AUC and AUPR are higher than 0.92 on both the benchmark datasets and combined datasets. It is noteworthy that the overall performance of the model is superior on the combined datasets compared to the benchmark datasets. For instance, among the benchmark datasets, the model on Dataset 1 and Dataset 2 achieves the AUC of 0.9514 and 0.9280, respectively. However, the model’s AUC on Dataset 4 (Dataset 1+Dataset 2) is 0.9690, which surpassed that of its constituent subsets. This improvement of performance on the combined dataset can be attributed to the more balanced relationship between the quantities of lncRNAs and drugs and the more trainable data in this dataset. However, on the contrary, the performance of benchmark Dataset 3 is better than that of the combined datasets. This discrepancy can be attributed to the presence of 4791 LDAs involving 4691 lncRNAs and 48 drugs in Dataset 3. Notably, the average degree size of each drug is 99, implying that one drug corresponds to multiple lncRNA information, enabling the model to extract more complex features from abundant lncRNA information. Consequently, it facilitates a more accurate capture of the relationship between drugs and lncRNAs. Overall, these observations indicate that the proposed method has an excellent performance on the five datasets.

www.frontiersin.org

Table 3. The performance of the method on five datasets.

3.3 Ablation experiment

To further study the influence of GCN and GAT on the model performance, we also conduct ablation experiments on five datasets. Specifically, we individually use the GCN and GAT modules, as well as their combined module for LDA prediction. As shown in Figure 3, among the three modules, the combined GCN and GAT modules obtain the optimal performance across all five datasets, followed by the GAT module, with the GCN module exhibiting the poorest performance. GCN module learns the feature representation of lncRNA and drug nodes by aggregating their neighbor information, which enables GCN to capture the spatial local structure of nodes. GAT module introduces an attention mechanism that allows the model to dynamically assign weights and integrate the global information of the lncRNA-drug bipartite graph, rather than just the neighbors of the lncRNA and drug nodes. Therefore, combining GCN and GAT modules enables the model to take full advantage of their strengths, complement each other, and further improve the prediction performance.

www.frontiersin.org

Figure 3. The results of the ablation experiment on five datasets.

3.4 Case study3.4.1 Predicting potential lncRNAs and drugs

In this section, we conduct a case study to further demonstrate the performance of the proposed method, we predict LDAs for two drugs (Berberine and Panobinostat) and two lncRNAs (NEAT1 and MEG3), respectively. First, we predict two drug-related lncRNAs using all known LDAs in Dataset 4 excluding those in Dataset 3 as the training data. Dataset 3 is used as the ground truth to test the predicted LDAs. Second, given that Dataset 1 contains more lncRNA information compared to drugs, two lncRNA-related drugs are predicted by employing all known LDAs in Dataset 1 except for those in Dataset 2 as the training data. Dataset 2 is the ground truth to test the predicted drugs. The top 10 predicted LDAs are ranked according to their prediction scores, among those, for any associations not shown in the test datasets (Dataset 3 and Dataset 2), we manually search relevant literature in PubMed to provide supporting evidence.

Figure 4 shows the top 10 predicted LDAs for the two drugs and two lncRNAs where the line width indicates the magnitude of the association score and the green lines indicate the confirmed LDAs in the test datasets. The blue lines indicate those LDAs that are not confirmed by the test dataset but have literature support. Generally, all the confirmed associations have large predicted scores. The red dotted lines represent LDAs having no support indication up to now. As shown in Figure 4A, 8 out of 10 predicted lncRNA-Berberine associations are validated, among which lncRNA “MALAT1” and “H19” are confirmed in the literature. Cao et al. (2020) demonstrated that lncRNA MALAT1 in cerebral ischemia was significantly reduced after treatment with the drug Berberine, highlighting its anti-inflammatory effects in mice after MCAO surgery. Song et al. (2022) identified lncRNA H19 as a potential key regulatory lncRNA of Berberine. Among the top 10 lncRNAs predicted associated with Panobinostat (Figure 4B), 8 lncRNAs are confirmed in test datasets. And for lncRNA NEAT1, 6 out of 10 predicted drugs related to NEAT1 are verified (Figure 4C). And 8 predicted drug-MEG3 associations have evidence in test datasets and literature (Figure 4D).

www.frontiersin.org

Figure 4. The predictive results of the top 10 lncRNAs associated with drugs Berberine (A) and Panobinostat (B), and the top 10 drugs related to lncRNA NEAT1 (C) and MEG3 (D). LncRNAs are represented by red nodes, while drugs are denoted as green nodes. The thickness of the edge indicates the predicted ranking, and the thicker the edge, the higher the ranking. Moreover, the LDAs found in test datasets are shown in green edges. Associations supported by existing literature are shown in blue edges, accompanied by their corresponding PMID references. Unconfirmed associations are represented by red dotted edges.

3.4.2 Functional enrichment analysis

Since one lncRNA may regulate the expression of multiple downstream genes, therefore intervening with one lncRNA may involve a variety of biological functions. In addition, a drug may target a certain lncRNA to alter its expression, thereby regulating the expression of lncRNA target genes or modifying the activity of signaling pathways to realize its biological functions. For instance, Guo et al. (2016) showed that aspirin could significantly induce the expression of lncRNA OLA1P2 in human colorectal cancer, thereby affecting the activity of the STAT3 signaling pathway. To gain insights into the functional implications of the drugs and lncRNAs of concern, we conduct functional enrichment analysis on their related genes by an online tool DAVID (Sherman et al., 2022), which is widely used for functional annotation and enrichment analysis of gene lists.

Based on the results of the first part of the case study, firstly, we perform functional enrichment analysis on the target genes of lncRNAs predicted associated with two drugs (Berberine and Panobinostat). We search the literature demonstrating that drug functions are related to the enrichment pathways of lncRNA target genes. In Figure 5A, among the top 15 KEGG pathways of Berberine-related lncRNA target genes, 12 have been confirmed associated with the existing functions of Berberine. For example, Ayati et al. (2017) demonstrated that Berberine can play an anticancer role by regulating the expression of oncomiRs and tumor-suppressive miRs in various cancer cells (Hepatocellular carcinoma, gastric cancer, ovarian cancer, colorectal cancer). Okuno et al. (2022) evidenced that Berberine overcomes gemcitabine-related chemical resistance by modulating rap1/PI3K-akt signaling in pancreatic ductal adenocarcinoma. In Figure 5B, there are 12 KEGG pathways of predicted Panobinostat-related lncRNA target genes that are found to be associated with the functions of Panobinostat. Lee et al. (2017) demonstrated that Panobinostat overcame resistance to gefitinib in KRAS-mutant/EGFR wild-type non-small-cell lung cancer by targeting TAZ. Studies have also shown that Panobinostat can restore the sensitivity of endocrine-resistant and triple-negative breast cancer cell lines to estrogen receptors (Tan et al., 2016).

www.frontiersin.org

Figure 5. The results of functio

留言 (0)

沒有登入
gif