Skip to main content

PlantNh-Kcr: a deep learning model for predicting non-histone crotonylation sites in plants

Abstract

Background

Lysine crotonylation (Kcr) is a crucial protein post-translational modification found in histone and non-histone proteins. It plays a pivotal role in regulating diverse biological processes in both animals and plants, including gene transcription and replication, cell metabolism and differentiation, as well as photosynthesis. Despite the significance of Kcr, detection of Kcr sites through biological experiments is often time-consuming, expensive, and only a fraction of crotonylated peptides can be identified. This reality highlights the need for efficient and rapid prediction of Kcr sites through computational methods. Currently, several machine learning models exist for predicting Kcr sites in humans, yet models tailored for plants are rare. Furthermore, no downloadable Kcr site predictors or datasets have been developed specifically for plants. To address this gap, it is imperative to integrate existing Kcr sites detected in plant experiments and establish a dedicated computational model for plants.

Results

Most plant Kcr sites are located on non-histones. In this study, we collected non-histone Kcr sites from five plants, including wheat, tabacum, rice, peanut, and papaya. We then conducted a comprehensive analysis of the amino acid distribution surrounding these sites. To develop a predictive model for plant non-histone Kcr sites, we combined a convolutional neural network (CNN), a bidirectional long short-term memory network (BiLSTM), and attention mechanism to build a deep learning model called PlantNh-Kcr. On both five-fold cross-validation and independent tests, PlantNh-Kcr outperformed multiple conventional machine learning models and other deep learning models. Furthermore, we conducted an analysis of species-specific effect on the PlantNh-Kcr model and found that a general model trained using data from multiple species outperforms species-specific models.

Conclusion

PlantNh-Kcr represents a valuable tool for predicting plant non-histone Kcr sites. We expect that this model will aid in addressing key challenges and tasks in the study of plant crotonylation sites.

Introduction

Post-translational modifications (PTMs) [1] of proteins involve the addition or removal of chemical groups to amino acid residues, thereby modifying protein properties and expanding functional diversity. PTMs play crucial roles in various biological processes and metabolic pathways. Among the PTMs, lysine crotonylation (Kcr) is a novel and significant modification that has been widely detected in both animals and plants. The Kcr modification was initially discovered in the histones of human somatic cells and mouse germ cells, and Kcr enrichment on sex chromosomes has been identified as a key indicator in male germ cell differentiation control [2]. In mice, it was shown that increased histone crotonylation level might have a positive effect on acute kidney injury [3]. In addition, histone Kcr sites are involved in many other biological processes, including organism development [2], DNA damage response [4], and gene transcription and expression [5]. Kcr is not limited to histones but is also abundant in non-histones [6,7,8,9]. Previous studies have clearly revealed that crotonylation of non-histone proteins is associated with various metabolic pathways and participates in protein expression and multiple cell signaling cascades [8].

In plants, global identification and functional analysis of lysine crotonylation have been conducted in species such as tabacum [10], papaya [11], rice [12], peanut [13], and wheat [14, 15]. The Kcr modification is involved in regulating various metabolic pathways in plants, including photosynthesis, oxidative phosphorylation, and carbon metabolism [14]. Additionally, Kcr is involved in gene transcription regulation [12] and adaptation to adverse conditions in plants [16]. Notably, Kcr is related to cold stress tolerance in plants [17], exhibiting a positive regulatory effect on wheat's freezing tolerance [14].

The current experimental methods for detecting Kcr sites include high-performance liquid chromatography fractionation, stable isotope labelling of amino acids in cell culture, immunological affinity enrichment, and high-resolution liquid chromatography-tandem mass spectrometry [18]. While biological experiments are the most reliable means to identify Kcr sites, the experiments are often time-consuming, labor-intensive, and costly. In addition, mass spectrometry platforms can only identify a subset of crotonylated peptides due to factors such as protein abundance, protein hydrolysis and digestion [19]. Therefore, computational models for conveniently and rapidly predicting Kcr sites are desirable, which have been developed in the past few years.

Early models for Kcr site prediction were limited by the small training datasets with fewer than 200 Kcr sites, and all sites were limited to histones. These models employed conventional machine learning methods such as support vector machines [20], random forest (RF) [21], LightGBM [22,23,24], etc. The input features used by these models include composition of amino acids and amino acid pairs, amino acid properties, etc. The representative models include CrotPred [25], CKSAAP_CrotSite [26], iKcr-PseEns [27], iCrotoK-PseAAC [28], LightGBM-CroSite [29], etc.

With the advancement of mass spectrometry technology, the global Kcr sites in the proteome of several species have been detected that enabled the utilization of significantly larger training datasets. At the same time, the deep learning frameworks [30] have reached a level of maturity that resulted in the common employment of deep learning methods in establishing predictive models for Kcr sites. Among these models, some predict Kcr sites on a mixture of histones and non-histones, such as Deep-Kcr [31], BERT-Kcr [32], DeepCap-Kcr [33], Adapt-Kcr [34], and ATCLSTM-Kcr [19]. Others are tailored to predict Kcr sites on non-histones, such as nhKcr [35], DeepKcrot [36], iKcr_CNN [37], and CapsNh-Kcr [38]. The primary input features of these models consist of binary encoding and embedding vectors. A majority of these models utilize convolutional neural networks (CNNs) [19, 31, 33,34,35,36,37,38] and long short-term memory networks (LSTM) [19, 32,33,34, 36] as integral components of their structure. Notably, some models integrate self-attention mechanism to enhance their predictive capabilities [19, 32, 34]. Additionally, a few models stand out due to their unique network architecture. For instance, DeepCap-Kcr and CapsNh-Kcr employ capsule networks, a distinctive approach in deep learning architecture. The deep learning models have resulted in significantly improved performance compared to the earlier conventional machine learning models.

Among the four models for predicting Kcr sites on non-histones, nhKcr, iKcr_CNN, and CapsNh-Kcr are limited to predicting human Kcr sites. In contrast, DeepKcrot predict Kcr sites in four species, including humans, rice, tabacum, and papaya. However, it is important to note that due to the current unavailability of DeepKcrot's web server, accessing its datasets and models for further research purposes can be challenging. In addition, recent experimental studies have detected Kcr sites in some other plants, emphasizing the need for a computational model that is specifically tailored for plants. To address this gap, it is essential to integrate existing Kcr site data detected from plants and establish a specialized computational model dedicated to plants.

In this study, we compiled a comprehensive dataset of non-histone Kcr sites from five plant species including rice, tabacum, papaya, peanut, and wheat, and built a reliable training and test dataset. Then we utilized the binary encoding (BE) as input features and employed a combination of a convolutional neural network (CNN) [39], a bidirectional long short-term memory (BiLSTM) network [40] and multi-head self-attention (MHSA) [41] to construct a novel deep learning model called PlantNh-Kcr. This model was specifically designed to predict Kcr sites on non-histones in plants. We validated our model through rigorous comparisons with conventional machine learning methods and other state-of-the-art deep learning models. On five-fold cross-validation and independent test, PlantNh-Kcr consistently demonstrated superior performance. Furthermore, it excelled in predicting Kcr sites across individual plant species, highlighting its remarkable versatility and generalizability. We believe that the development of this plant-specific prediction model offers valuable insights for the biological community and will drive further advancement in plant biology.

Materials and methods

Benchmark dataset

To train and test the model, we carefully curated a training dataset and a test dataset. This process was meticulously designed and is depicted in Fig. 1. We first collected non-histone Kcr sites from five plant species. These sites numbered 5692 from wheat [14, 15], 1258 from rice [12], 2028 from tabacum [10], 6603 from peanut [13], and 5332 from papaya [11]. We then retrieved the corresponding protein sequences from the UniProt [42] and NCBI [43] databases for each species. Subsequently, we extracted peptides of length 29 from these protein sequences, with the K (Lysine) residue positioned at the center and 14 residues upstream and downstream respectively. If a peptide had fewer than 14 residues on one side, we replaced the missing residues with X. Peptides where the central K residues represented Kcr sites were designated as positive samples; otherwise, they were designated as negative samples. To eliminate redundancy and potential false negatives, we used the CD-HIT program [44] with a sequence identity threshold of 40%. Finally, we obtained 12,352 positive and 46,389 negative samples. To evaluate the performance of the model on test samples of each species, we separated the samples based on their species. For each species, the samples were randomly divided into two subsets in a 7:3 ratio while retaining the proportion of positive and negative samples during the partitioning process. The larger sets for each species were merged to form the training dataset and the smaller sets for each species were merged to form the test dataset. The training dataset totaled 41,114 samples, with 8644 positive and 32,470 negative samples. The test dataset included 17,627 samples, with 3708 positive 13,919 negative samples. The numbers of samples for each species in the training and test datasets are listed in Table 1.

Fig. 1
figure 1

The flowchart of dataset preparation

Table 1 The numbers of positive and negative samples for each species in the training and test datasets

Peptide encoding

To build a predictive model for non-histone Kcr sites, it is necessary to transform peptide samples into numerical vectors as input features for the model. PlantNh-Kcr employs binary encoding as its input features. We conducted a comparative analysis of PlantNh-Kcr’s predictive performance against conventional machine learning models and other deep learning models. The conventional machine learning models utilize various input features, including amino acid composition (AAC), enhanced group amino acid composition (EGAAC), BE, AAindex encoding, and BLOSUM62 encoding. Other deep learning models employ features including BE, word embedding (WE) encoding, AAindex encoding, and BLOSUM62 encoding. The following provides a detailed description of these encoding methods:

AAC: In bioinformatics, AAC is a commonly used encoding method, which calculates the frequencies of each amino acid in a peptide. In this study, X is also considered as an amino acid. So the peptide is encoded as a 21-dimensional vector, where each dimension corresponds to the frequency of one of the 21 amino acids present in the peptide.

EGAAC: The EGAAC encoding divides amino acids into five groups based on their physicochemical properties, i.e. aliphatic group (GAVLMI), aromatic group (FYW), positively charged group (KRH), negatively charged group (DE), and no charge group (STCPNQ). A peptide is encoded as a five-dimensional vector, where each dimension represents the proportion of one of the five groups of amino acids within the peptide.

BE: BE is a common technique used to convert amino acid sequences into numerical representations suitable for model training. For this encoding method, each amino acid is encoded as a 21-dimensional binary vector. This vector has one component set to 1 to indicate the type of amino acid, while all other components are set to 0. Finally, a peptide of length 29, is encoded as a matrix or a vector of size 29 \(\times\) 21.

WE encoding: WE is a technique that has gained popularity in the field of natural language processing. It assigns words to vectors in a high-dimensional space, ensuring that semantically similar words are positioned close to each other. This technique has also been effectively applied to sequence encoding in bioinformatics [45, 46]. In this work, the vocabulary size is set to 21, representing the number of amino acid types. The peptide of length 29 is treated as a sentence, with each amino acid residue mapped to a unique word ID. Subsequently, WE is used to translate these IDs into vectors. Finally, the peptide is encoded as a matrix of size 29 \(\times\) 10, where 10 is the dimension of the vector space.

AAindex encoding: AAindex [47] is a public database that curates a range of physicochemical and biochemical properties of amino acids. This database serves as a valuable resource for various bioinformatics studies, including protein structure prediction, sequence alignment, protein function annotation, and more. Previously, the model nhKcr selected 29 indices from AAindex that were most relevant to the prediction task to encode the peptide [35]. In this study, we used the same 29 indices to encode the peptide. Consequently, the peptide of length 29 is encoded as a matrix or a vector of size 29 \(\times\) 29.

BLOSUM62 encoding: BLOSUM62 (BLOck Substitution Matrix 62) [48], is a widely used substitution matrix in protein sequence alignment. The matrix assigns scores to pairs of amino acids based on their substitution frequency during evolution. Higher scores indicate more frequent substitutions, while lower scores indicate rare substitutions. In this study, we used the rows of the BLUSUM62 matrix to encode amino acids in the peptide. Consequently, the peptide of length 29 is encoded as a matrix or a vector of size 29 \(\times\) 21.

The structures of the plantNh-Kcr model

The structure of PlantNh-Kcr was determined as Fig. 2 after evaluating various encoding methods and model architectures. The model accepts a 29 × 21 matrix derived from binary encoding as input. This matrix feeds into two distinct layers. The first is a convolutional layer that is followed by two additional convolutional layers. The second layer is a BiLSTM layer that is succeeded by a MHSA layer. The outputs of the third convolutional layer and the MHSA layer are merged and flattened into a vector. The flatten layer is followed by a linear layer and an output layer. All the layers are described in detail below.

Fig. 2
figure 2

The architecture of the PlantNh-Kcr model

Input layer: The layer receives a 29 \(\times\) 21 matrix as input.

Convolutional layers: The first convolutional layer has 21 input channels and 32 output channels, with a kernel size of 5 and a stride of 1. The second convolutional layer has 32 input channels and 32 output channels, and the third one has 32 input channels and 29 output channels. Both the latter two layers have a kernel size of 5 and a stride of 2. The outputs of each layer are activated using the ReLU function [49]. During training, to prevent overfitting, 30% of the output data from the three convolution layers are dropped respectively.

BiLSTM layer and MHSA layer: The input size of the BiLSTM layer is 21, and the output size is 128. The MHSA layer has an input size of 128 and eight attention heads. To prevent overfitting, dropout operations with ratios of 0.9 and 0.5 are applied after the BiLSTM layer and MHSA layer, respectively.

Flatten layer: The flatten layer flattens the concatenated outputs of the third convolutional layer and MHSA layer, resulting in a 3944-dimensional vector.

Linear layer: The input size of the linear layer is 3944 and the output size is 128. The output is activated using the ReLU function.

Output layer: The output layer has an input size of 128 and output size of 2. The two-dimensional output vector represents the probabilities of a sample being positive and negative, respectively.

Focal loss

In this study, the training dataset has significantly more negative samples than positive samples, which would lead to a bias towards the negative samples during model training. To address this issue, we employed focal loss [50] as the loss function for the model. Focal loss reshapes the traditional cross-entropy loss function by introducing a modulating factor \({(1-{p}_{t})}^{\gamma }\). The mathematical formulas for focal loss are as follows:

$${p}_{t}=\left\{\begin{array}{l}p \qquad \quad if\, y=1\\ 1-p \quad \,otherwise.\end{array}\right. 0\le p\le 1$$
(1)
$${\alpha }_{t}=\left\{\begin{array}{l}\alpha \qquad \quad if\, y=1\\ 1-\alpha \quad \,otherwise.\end{array}\right. 0\le \alpha \le 1$$
(2)
$${\text{FL}}\left({p}_{t}\right)=-{\alpha }_{t}{\left(1-{p}_{t}\right)}^{\gamma }log({p}_{t}) \quad 0\le {\alpha }_{t}\le 1,\gamma \ge 0$$
(3)

where \(p\) represents the probability for the sample to be class 1 (1 represents positive samples, while 0 represents negative samples). \({\alpha }_{t}\) represents the balanced weight factor for the sample. The modulating factor \({(1-{p}_{t})}^{\gamma }\) is used to adjust the weight of easy samples and hard samples, and \(\upgamma\) indicates a tunable focusing parameter.

Optimization of the model

The PlantNh-Kcr model was constructed and trained in a Python 3.9 and Pytorch 1.13.1 environment. Focal loss [50], with α set to 0.7 and γ to 1, was employed as the loss function. The model optimization was achieved using the Adam algorithm [51] with a learning rate of 0.001. The batch size of the input data during training was set to 256, and the number of training epochs was set to 50. To determine the optimal hyperparameters for the model, grid search was employed.

Model evaluation

In bioinformatics, the evaluation of classification models often involves cross-validation and independent test to assess their generalization capabilities. Similar to previous studies [33, 35, 37, 38] that predicted Kcr sites, we employed five-fold cross-validation and independent tests to evaluate PlantNh-Kcr and other models. For this purpose, we prepared the training dataset and the test dataset. For five-fold cross-validation, the training dataset was evenly divided into five folds. Four folds were used to learn the model, while the remaining one was used to validate its performance. This process was repeated five times, ensuring that each fold was used once for validation. For independent test, the training dataset was used to build the model, and the test dataset was used to access its performance.

In bioinformatics, commonly used evaluation metrics for binary classification models include sensitivity (Sn), specificity (Sp), accuracy (ACC), F1-score, Matthews correlation coefficient (MCC), and area under the receiver operating characteristic (ROC) curve (AUC) [35, 38, 52, 53]. The mathematical formulas for Sn, Sp, ACC, and MCC are as follows:

$${\text{Sn}}=\frac{{\text{TP}}}{{\text{TP}}+{\text{FN}}}$$
(4)
$${\text{Sp}}=\frac{{\text{TN}}}{{\text{TN}}+{\text{FP}}}$$
(5)
$${\text{ACC}}=\frac{{\text{TP}}+{\text{TN}}}{{\text{TP}}+{\text{TN}}+{\text{FP}}+{\text{FN}}}$$
(6)
$${\text{F}}1-{\text{score}}=\frac{2\times {\text{TP}}}{2\times {\text{TP}}+{\text{FP}}+{\text{FN}}}$$
(7)
$${\text{MCC}}=\frac{\left({\text{TP}}\times {\text{TN}}\right)-\left({\text{FP}}\times {\text{FN}}\right)}{\sqrt{\left({\text{TP}}+{\text{FP}}\right)\left({\text{TP}}+{\text{FN}}\right)\left({\text{TN}}+{\text{FP}}\right)\left({\text{TN}}+{\text{FN}}\right)}}$$
(8)

In the above equations, TP, FP, TN, and FN represent the numbers of true positives, false positives, true negatives, and false negatives, respectively. Sn indicates the ability of the model to identify positive samples, with higher values indicating more accurate predictions for positive samples. Sp reflects the ability of the model to identify negative samples, with higher values indicating more accurate predictions for negative samples. F1-score provides a comprehensive measure of the model's performance in identifying positive samples, through balancing the counts of true positives, false positives, and false negatives. A higher F1-score indicates better performance. MCC considers both Sn and Sp, and ranges from − 1 to 1. A higher MCC value indicates better performance of the model. The ROC curve offers a graphical representation of the relationship between the true positive rate (TPR) and false positive rate (FPR) at different thresholds. TPR corresponds to Sn, while FPR equals one minus Sp. AUC represents the probability of a model ranking positive samples above negative samples. AUC is regarded as the most important metric in the evaluation of many bioinformatics models. The closer the ROC curve approaches the upper left corner, the closer the AUC value approaches 1, indicating a better classification performance of the model. In this study, samples with a predicted probability of being positive greater than 0.5 are classified as positive samples. The evaluation metrics of Sn, Sp, ACC, F1-score, and MCC are computed based on the fixed threshold of 0.5. We primarily use the ROC curve and its corresponding AUC value to compare the performance of different models. The ROC curve effectively visualizes the trade-off between Sn and Sp across various thresholds. This allows to compare models by considering their respective sensitivities at the same specificities, thus providing a more comprehensive evaluation.

To ensure the robustness of our model's performance, we conducted rigorous tests. For five-fold cross-validation, we calculated the mean and standard deviation of the metric values obtained from each fold. For independent test, we conducted 10 independent tests with different random seeds, and calculated the mean and standard deviation of the evaluated metrics.

Results

Conservation analysis of non-histone Kcr sites in plants

Kcr is a post-translational modification that plays a crucial role in various cellular processes. It has been observed that the evolution of Kcr sites exhibits conservation, which suggests that these sites have functional significance [35]. To further investigate the conservation of plant non-histone Kcr sites, we used the pLogo tool [54] to generate a sequence logo (Fig. 3A) using the merged training and test dataset. Significant disparities can be observed in the distribution of some amino acids such as K, D, E, R, and P, surrounding Kcr sites and non-Kcr sites. Notably, residues D and E are overrepresented at positions -1 and + 1. Residue K is prevalent at numerous positions, and it is more overrepresented on the left side of Kcr sites. Residues R and P are significantly underrepresented at positions − 1 and + 1, respectively. Previous research has identified specific motifs that are enriched around Kcr sites, including “EkxxxxxK”, “EkxxxK”, and “KxxxEK”, where x denotes any amino acid [35, 38]. These findings are consistent with our sequence logo, which indicates an overrepresentation of amino acid residues K and E in the vicinity of Kcr sites in plants. To further contextualize our findings, we also generated a sequence logo (Fig. 3B) for Kcr sites on non-histone proteins in humans. The resulting logo was based on the training dataset of nhKcr, a model developed for predicting Kcr sites on human non-histones [35]. Notably, the amino acid distribution observed around human Kcr sites exhibits similarity to that in plants; however, there were slight differences. For example, K is underrepresented at positions − 1 and + 1 for the human Kcr sites, which is not observed in plants. This observation highlights the need for developing a predictive model dedicated to plants.

Fig. 3
figure 3

Sequence logo of Kcr sites on non-histone proteins. A Sequence logo for plants; B Sequence logo for humans

Performance of PlantNh-Kcr on five-fold cross-validation and independent tests

To evaluate the performance of PlantNh-Kcr, we conducted a five-fold cross-validation and 10 independent tests. For cross-validation, the ROC curves for each fold were tightly clustered in the top left corner of the plot (Fig. 4A), indicating that the model has strong discriminatory power. The average AUC value are 0.891, which is significantly better than random prediction. The average values for Sn, Sp, ACC, F1-score and MCC are 0.821, 0.810, 0.812, 0.648 and 0.551 respectively (Table 2). For independent tests, PlantNh-Kcr also demonstrats strong performance, with the average AUC value of 0.899 (Fig. 4B) and the average values for Sn, Sp, ACC, F1-score, and MCC are 0.811, 0.833, 0.828, 0.665 and 0.572, respectively (Table 2).

Fig. 4
figure 4

ROC curves of the PlantNh-Kcr model on five-fold cross-validation and independent tests. A The ROC curves on five-fold cross-validation; B The ROC curve on independent tests

Table 2 Metric values of the PlantNh-Kcr model on five-fold cross-validation and independent tests

To visualize the discriminatory power of PlantNh-Kcr, we utilized the training dataset to train a model and subsequently fed the samples in the test dataset to it. Then we used t-SNE [55] for dimensionality reduction and visualization of the input data in the input layer, the output data of the flatten layer, and the output data of the linear layer. The results are presented in Fig. 5. In this figure, the red and light blue dots represent Kcr and non-Kcr sites, respectively. It is evident from the input layer that Kcr and non-Kcr sites are intermingled. However, following the processing involving three convolutional layers, a BiLSTM layer, and a MHSA layer, most Kcr and non-Kcr sites are separated. This observation underscores the discriminatory capability of these layers in effectively distinguishing between Kcr and non-Kcr sites. Subsequently, after the linear layer, Kcr sites are predominantly clustered in the left region, forming a distinct boundary from non-Kcr sites. This outcome highlights the model’s prowess in accurately classifying Kcr sites.

Fig. 5
figure 5

T-SNE visualization of test samples in PlantNh-Kcr layers. A The input layer; B The flatten layer; C The linear layer

Comparison with conventional machine learning models and deep learning models

To further demonstrate the superior performance and robustness of PlantNh-Kcr, we conducted a comparative analysis with several well-established conventional machine learning models and deep learning models. The detailed information about these models is provided in Additional file 1.

In this study, we utilized three conventional machine learning methods including RF [21], AdaBoost [56], and LightGBM [24]. Additionally, five commonly used encodings were employed, including BE, ACC, EGAAC, AAindex, and BLOUSUM62. Each encoding was fed into each conventional machine learning model for training. The specific results for each combination are summarized in Tables 3, 4. For both five-fold cross-validation and independent tests, RF, AdaBoost, and LightGBM get the largest AUC values, when BE, AAindex, and BLOSUM62 encodings were used as input features, respectively. Among the three models, the LightGBM model emerged as the leader with average AUC values of 0.869 and 0.881 on five-fold cross-validation and independent tests, respectively. However, it lagged behind PlantNh-Kcr in terms of performance. To visually compare the models, Fig. 6 presents the ROC curves on independent tests. Notably, The ROC curve of PlantNh-Kcr is above those of other models. This observation indicates that at the same FPR, PlantNh-Kcr exhibits the highest TPR, indicating its superior ability to correctly identify positive samples compared to other models when predicting the same number of false positives.

Table 3 Metric values of different models on five-fold cross-validation
Table 4 Metric values of different models on independent test
Fig. 6
figure 6

ROC curves of different models on independent tests

Three networks including LSTM network, BiLSTM network, and CNN were used to compare with PlantNh-Kcr. The inputs for these networks encompassed BE, WE, AAindex, and BLOSUM62 encodings. The specific metric values for each network are detailed in Tables 3,  4. Interestingly, all three networks perform best when using BE as input. On five-fold cross-validation, the maximum average AUC values achieved by the LSTM, BiLSTM, and CNN networks are 0.882, 0.880 and 0.888, respectively. Similarly, on independent tests, the maximum average AUC values of these networks are 0.890, 0.889, and 0.896, respectively. However, the performance of the three networks is still inferior to PlantNh-Kcr.

Comparison with existing models for Kcr site prediction on non-histones

To further demonstrate the performance of our model, we conducted a comparative analysis with four other models: nhKcr, iKcr_CNN, CapsNh-Kcr, and DeepKcrot, all designed to predict Kcr sites on non-histones. nhKcr, iKcr_CNN and CapsNh-Kcr predict Kcr sites in human. The nhKcr model integrated BE, AAindex encoding and BLOSUM62 encoding as input features and employed a CNN architecture. The iKcr_CNN model employed a CNN architecture and utilized a focal loss function for optimization. CapsNh-Kcr employed a CNN-based capsule network strategy. DeepKcrot predicted Kcr sites in four species including human, rice, papaya and tabacum. It utilized CNN with WE encoding as input features.

For nhKcr, iKcr_CNN and CapsNh-Kcr, we downloaded their source codes. For DeepKcrot, we rewrote its code due to the unavailability of its web server. We applied focal loss to nhKcr and DeepKcrot because they didn’t address the data imbalance issue in their original source codes. We then trained the four models using the training dataset and evaluated their performance on the test set. The prediction performance of the four models is shown in Fig. 7 and Table 5. The average AUC values are 0.876, 0.876, 0.890, and 0.892, respectively, which are lower than that of PlantNh-Kcr. This again underscores the superior performance of PlantNh-Kcr.

Fig. 7
figure 7

ROC curves of PlantNh-Kcr and the other four models

Table 5 Metric values of PlantNh-Kcr and the other four models

Ablation study

To assess the effect of each component in the PlantNh-Kcr model on prediction performance, we conducted an ablation study. In this study, we removed the linear layer, CNN, MHSA, and BiLSTM + MHSA individually from the model and evaluated the prediction performance on independent tests. The results are summarized in Table 6. Removing the linear layer and CNN individually resulted in a decrease of 1.1% and 1.2% in AUC values, respectively. This suggests that these two components have a certain impact on the overall performance of the model. On the other hand, removing MHSA and BiLSTM + MHSA individually resulted in a decrease of 0.5% and 0.3% in AUC values, respectively, indicating that these components have a smaller impact on performance compared to the linear layer and CNN. Overall, our results demonstrate that each component in the PlantNh-Kcr model contributes to its prediction performance. Removing any module from the model will result in a decrease in performance, indicating that each module is essential for achieving optimal performance.

Table 6 Prediction performance of models in the ablation study

The performance of PlantNh-Kcr on independent tests for each plant

In this study, we collected non-histone Kcr sites from different types of plants. Given the potential species-specific impact on these sites, it’s necessary to assess the generalizability of our predictive model across diverse plant species. Therefore, we studied the performance of our model for each species on independent tests. Table 7 details the evaluation metrics for each species, which are further visualized in Fig. 8 as a bar chart.

Table 7 Performance of PlantNh-Kcr for different plants
Fig. 8
figure 8

Metric values on independent tests for different plants

The results indicate that the prediction performance of the model varies slightly across different species. Notably, peanuts and papaya exhibit particularly strong performance, with average AUC values of 0.930 and 0.914, respectively. The model also demonstrates good performance for tabacum and rice, with average AUC values of 0.896 and 0.890, respectively. However, wheat exhibits slightly lower performance compared to other species, with an average AUC value of 0.858. This may be attributed to species-specific characteristics.

To study the performance of species-specific models, we developed individual models for each plant using samples from the corresponding species in the training dataset. We then evaluated these models using samples from the corresponding species in the test set. The performance of these models on five metrics are shown in Table 8. Notably, the peanut-specific and papaya-specific models exhibit the best performance, with average AUC values of 0.920 and 0.902, respectively. In contrast, the species-specific models for rice, tabacum, and wheat exhibit relatively poorer performance. This can be attributed to the smaller training set size for rice and tabacum and potential species-specific characteristics affecting crotonylation patterns in wheat. When compared with the general model’s performance in Table 7, the species-specific models underperform. This finding underscores the advantage of integrating data from diverse species to train a general predictive model for plant non-histone Kcr sites.

Table 8 Performance of species-specific models on independent tests

Discussion

The PlantNh-Kcr exhibits superior performance. However, there are still some issues that need to be considered.

First, our model PlantNh-Kcr contains three convolutional layers, which can effectively capture local patterns in protein sequences. Careful consideration must be given to the kernel size and the step size, as well as the number of convolution kernels. Too few or too many convolution kernels can lead to information loss or overfitting, respectively, which can impact model performance. Furthermore, when utilizing the convolutional layer to process long protein sequences, there is a risk of losing global contextual information. This can be a limiting factor in the predictive accuracy of the model. To address this issue, stacking multiple convolutional layers and effectively integrating their outputs can compensate for the loss of global context. By doing so, the model can achieve a more comprehensive understanding of the protein sequences, ultimately leading to improved predictive performance.

Second, multiple encodings were described in the paper, such as BE, WE encoding, AAindex encoding, and BLOSUM62 encoding. The PlantNh-Kcr model only utilize BE as input features. We have attempted to integrate multiple encodings as input features of the model, but failed to improve the performance. This may be because these features have poor complementarity.

Third, there are far more negative samples than positive samples in our training set. This imbalance can significantly influence model training, biasing it towards the negative samples. To address this issue, three methods were employed: up-sampling the positive samples, down-sampling the negative samples, and utilizing the focal loss function. Among these methods, the focal loss function presented the best prediction performance, and improved the ability of the model to correctly predict positive samples. We believe that dataset imbalance remains a potential problem that needs to be addressed in bioinformatics.

Conclusion

In this study, we compiled a large dataset of non-histone Kcr sites from five different plant species. Using this dataset, we developed a deep learning model called PlantNh-Kcr to predict non-histone Kcr sites in plants. The model’s architecture integrates CNN, LSTM, and attention mechanism, utilizing BE as its primary input features. Notably, the model exhibits satisfactory performance on both five-fold cross-validation and independent tests, outperforming several other models. In addition, there are minor variations in prediction performance across different plant species, a general predictive model demonstrates superior performance compared to species-specific models. We believe that the PlantNh-Kcr model offers a valuable contribution to addressing challenges and advancing the study of plant Kcr sites. We also believe that as more Kcr sites are experimentally determined and as deep learning techniques continue to develop, we will see the emergence of more high-performance models for predicting Kcr sites.

Availability of data and materials

The source code and datasets are publicly available at https://github.com/jiangyanming-individual/PlantNh-Kcr.

References

  1. Bao W, Yang B, Chen B. 2-hydr_ensemble: lysine 2-hydroxyisobutyrylation identification with ensemble method. Chemometr Intell Laboratory Syst. 2021. https://doi.org/10.1016/j.chemolab.2021.104351.

    Article  Google Scholar 

  2. Tan M, Luo H, Lee S, et al. Identification of 67 Histone marks and histone lysine crotonylation as a new type of histone modification. Cell. 2011;146:1016–28. https://doi.org/10.1016/j.cell.2011.08.008.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  3. Ruiz-Andres O, Sanchez-Niño MD, Cannata-Ortiz P, et al. Histone lysine-crotonylation in acute kidney injury. Dis Model Mech. 2016. https://doi.org/10.1242/dmm.024455.

    Article  PubMed  PubMed Central  Google Scholar 

  4. Abu-Zhayia ER, Machour FE, Ayoub N. HDAC-dependent decrease in histone crotonylation during DNA damage. J Mol Cell Biol. 2019;11:804–6. https://doi.org/10.1093/jmcb/mjz019.

    Article  PubMed  PubMed Central  Google Scholar 

  5. Montellier E, Rousseaux S, Zhao Y, et al. Histone crotonylation specifically marks the haploid male germ cell gene expression program: post-meiotic male-specific gene expression. BioEssays. 2012;34:187–93. https://doi.org/10.1002/bies.201100141.

    Article  CAS  PubMed  Google Scholar 

  6. Wu Q, Li W, Wang C, et al. Ultradeep lysine crotonylome reveals the crotonylation enhancement on both histones and nonhistone proteins by SAHA treatment. J Proteome Res. 2017;16:3664–71. https://doi.org/10.1021/acs.jproteome.7b00380.

    Article  CAS  PubMed  Google Scholar 

  7. Wei W, Mao A, Tang B, et al. Large-scale identification of protein crotonylation reveals its role in multiple cellular functions. J Proteome Res. 2017;16:1743–52. https://doi.org/10.1021/acs.jproteome.7b00012.

    Article  CAS  PubMed  Google Scholar 

  8. Xu W, Wan J, Zhan J, et al. Global profiling of crotonylation on non-histone proteins. Cell Res. 2017;27:946–9. https://doi.org/10.1038/cr.2017.60.

    Article  PubMed  PubMed Central  Google Scholar 

  9. Hou JY, Zhou L, Li JL, et al. Emerging roles of non-histone protein crotonylation in biomedicine. Cell Biosci. 2021;11:101. https://doi.org/10.1186/s13578-021-00616-2.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  10. Sun H, Liu X, Li F, et al. First comprehensive proteome analysis of lysine crotonylation in seedling leaves of Nicotiana tabacum. Sci Rep. 2017;7:3013. https://doi.org/10.1038/s41598-017-03369-6.

    Article  CAS  PubMed  PubMed Central  ADS  Google Scholar 

  11. Liu K, Yuan C, Li H, et al. A qualitative proteome-wide lysine crotonylation profiling of papaya (Carica papaya L.). Sci Rep. 2018;8:8230. https://doi.org/10.1038/s41598-018-26676-y.

    Article  CAS  PubMed  PubMed Central  ADS  Google Scholar 

  12. Liu S, Xue C, Fang Y, et al. Global involvement of lysine crotonylation in protein modification and transcription regulation in rice. Mol Cell Proteomics. 2018;17:1922–36. https://doi.org/10.1074/mcp.RA118.000640.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  13. Xu M, Luo J, Li Y, et al. First comprehensive proteomics analysis of lysine crotonylation in leaves of peanut (Arachis hypogaea L.). Proteomics. 2021;21:e2000156. https://doi.org/10.1002/pmic.202000156.

    Article  CAS  PubMed  Google Scholar 

  14. Zhang N, Wang S, Zhao S, et al. Global crotonylatome and GWAS revealed a TaSRT1-TaPGK model regulating wheat cold tolerance through mediating pyruvate. Sci Adv. 2023;9:eadg1012. https://doi.org/10.1126/sciadv.adg1012.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  15. Zhu D, Liu J, Duan W, et al. Analysis of the chloroplast crotonylome of wheat seedling leaves reveals the roles of crotonylated proteins involved in salt-stress responses. J Exp Bot. 2023;74:2067–82. https://doi.org/10.1093/jxb/erad006.

    Article  CAS  PubMed  Google Scholar 

  16. Lu Y, Xu Q, Liu Y, et al. Dynamics and functional interplay of histone lysine butyrylation, crotonylation, and acetylation in rice under starvation and submergence. Genome Biol. 2018;19:144. https://doi.org/10.1186/s13059-018-1533-y.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  17. Lin P, Bai HR, He L, et al. Proteome-wide and lysine crotonylation profiling reveals the importance of crotonylation in chrysanthemum (Dendranthema grandiforum) under low-temperature. BMC Genomics. 2021;22:51. https://doi.org/10.1186/s12864-020-07365-5.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  18. Yu H, Bu C, Liu Y, et al. Global crotonylome reveals CDYL-regulated RPA1 crotonylation in homologous recombination-mediated DNA repair. Sci Adv. 2020;6:eaay4697. https://doi.org/10.1126/sciadv.aay4697.

    Article  CAS  PubMed  PubMed Central  ADS  Google Scholar 

  19. Yang YH, Wu SF, Kong J, et al. Using ATCLSTM-Kcr to predict and generate the human lysine crotonylation database. J Proteomics. 2023;281: 104905. https://doi.org/10.1016/j.jprot.2023.104905.

    Article  CAS  PubMed  Google Scholar 

  20. Joachims T. Making large-scale SVM learning practical. Technical report, 1998.

  21. Breiman LJMI. Random forests. Mach Learn. 2001;45:5–32.

    Article  Google Scholar 

  22. Bao W, Cui Q, Chen B, et al. Phage_UniR_LGBM: phage virion proteins classification with UniRep features and lightGBM model. Comput Math Methods Med. 2022;2022:9470683. https://doi.org/10.1155/2022/9470683.

    Article  PubMed  PubMed Central  Google Scholar 

  23. Bao W, Gu Y, Chen B, et al. Golgi_DF: golgi proteins classification with deep forest. Front Neurosci. 2023;17:1197824. https://doi.org/10.3389/fnins.2023.1197824.

    Article  PubMed  PubMed Central  Google Scholar 

  24. Ke G, Meng Q, Finley T, et al. Lightgbm: a highly efficient gradient boosting decision tree. Front Neurosci. 2017. https://doi.org/10.3389/fnins.2023.1197824.

    Article  Google Scholar 

  25. Huang G, Zeng W. A discrete hidden Markov model for detecting histone crotonyllysine sites. J Mol Graph Modell. 2016;75:717–30.

    MathSciNet  Google Scholar 

  26. Ju Z, He JJ. Prediction of lysine crotonylation sites by incorporating the composition of k-spaced amino acid pairs into Chou’s general PseAAC. J Mol Graph Model. 2017;77:200–4. https://doi.org/10.1016/j.jmgm.2017.08.020.

    Article  CAS  PubMed  Google Scholar 

  27. Qiu WR, Sun BQ, Xiao X, et al. iKcr-PseEns: Identify lysine crotonylation sites in histone proteins with pseudo components and ensemble classifier. Genomics. 2018;110:239–46. https://doi.org/10.1016/j.ygeno.2017.10.008.

    Article  CAS  PubMed  Google Scholar 

  28. Malebary SJ, Rehman MSU, Khan YD. iCrotoK-PseAAC: Identify lysine crotonylation sites by blending position relative statistical features according to the Chou’s 5-step rule. PLoS ONE. 2019;14: e0223993. https://doi.org/10.1371/journal.pone.0223993.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  29. Liu Y, Yu Z, Chen C, et al. Prediction of protein crotonylation sites through LightGBM classifier based on SMOTE and elastic net. Anal Biochem. 2020;609: 113903. https://doi.org/10.1016/j.ab.2020.113903.

    Article  CAS  PubMed  Google Scholar 

  30. Meng R, Yin S, Sun J, et al. scAAGA: Single cell data analysis framework using asymmetric autoencoder with gene attention. Comput Biol Med. 2023;165: 107414. https://doi.org/10.1016/j.compbiomed.2023.107414.

    Article  CAS  PubMed  Google Scholar 

  31. Lv H, Dao F-Y, Guan Z-X, et al. Deep-Kcr: accurate detection of lysine crotonylation sites using deep learning method. Briefings Bioinf. 2021. https://doi.org/10.1093/bib/bbaa255.

    Article  Google Scholar 

  32. Qiao Y, Zhu X, Gong H, et al. BERT-Kcr: prediction of lysine crotonylation sites by a transfer learning method with pre-trained BERT models. Bioinformatics. 2022;38:648–54. https://doi.org/10.1093/bioinformatics/btab712.

    Article  CAS  PubMed  Google Scholar 

  33. Khanal J, Tayara H, Zou Q, et al. DeepCap-Kcr: accurate identification and investigation of protein lysine crotonylation sites based on capsule network. Brief Bioinform. 2022. https://doi.org/10.1093/bib/bbab492.

    Article  PubMed  Google Scholar 

  34. Li Z, Fang J, Wang S, et al. Adapt-Kcr: a novel deep learning framework for accurate prediction of lysine crotonylation sites based on learning embedding features and attention architecture. Brief Bioinform. 2022. https://doi.org/10.1093/bib/bbac037.

    Article  PubMed  PubMed Central  Google Scholar 

  35. Chen YZ, Wang ZZ, Wang Y, et al. nhKcr: a new bioinformatics tool for predicting crotonylation sites on human nonhistone proteins based on deep learning. Brief Bioinform. 2021. https://doi.org/10.1093/bib/bbab146.

    Article  PubMed  PubMed Central  Google Scholar 

  36. Wei X, Sha Y, Zhao Y, et al. DeepKcrot: a deep-learning architecture for general and species-specific lysine crotonylation site prediction. IEEE Access. 2021;9:49504–13. https://doi.org/10.1109/access.2021.3068413.

    Article  Google Scholar 

  37. Dou L, Zhang Z, Xu L, et al. iKcr_CNN: A novel computational tool for imbalance classification of human nonhistone crotonylation sites based on convolutional neural networks with focal loss. Comput Struct Biotechnol J. 2022;20:3268–79. https://doi.org/10.1016/j.csbj.2022.06.032.

    Article  PubMed  PubMed Central  Google Scholar 

  38. Khanal J, Kandel J, Tayara H, et al. CapsNh-Kcr: Capsule network-based prediction of lysine crotonylation sites in human non-histone proteins. Comput Struct Biotechnol J. 2023;21:120–7. https://doi.org/10.1016/j.csbj.2022.11.056.

    Article  CAS  PubMed  Google Scholar 

  39. Kim YJapa. Convolutional neural networks for sentence classification 2014.

  40. Schuster M, Paliwal KK. Bidirectional recurrent neural networks. IEEE Trans Signal Process. 1997;45:2673–81. https://doi.org/10.1109/78.650093.

    Article  ADS  Google Scholar 

  41. Vaswani A, Shazeer N, Parmar N, et al. Attention is all you need. Nucleic Acids Res. 2017. https://doi.org/10.1093/nar/gkr1048.

    Article  Google Scholar 

  42. Dimmer EC, Huntley RP, Alam-Faruque Y, et al. The UniProt-GO annotation database in 2011. Nucleic Acids Res. 2012;40:D565-570. https://doi.org/10.1093/nar/gkr1048.

    Article  CAS  PubMed  Google Scholar 

  43. Benson DA, Karsch-Mizrachi I, Lipman DJ, et al. GenBank. Nucleic Acids Res. 2009;37:D26-31. https://doi.org/10.1093/nar/gkn723.

    Article  CAS  PubMed  Google Scholar 

  44. Huang Y, Niu B, Gao Y, et al. CD-HIT Suite: a web server for clustering and comparing biological sequences. Bioinformatics. 2010;26:680–2. https://doi.org/10.1093/bioinformatics/btq003.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  45. Yang KK, Wu Z, Bedbrook CN, et al. Learned protein embeddings for machine learning. Bioinformatics. 2018;34:2642–8. https://doi.org/10.1093/bioinformatics/bty178.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  46. Lyu X, Li S, Jiang C, et al. DeepCSO: a deep-learning network approach to predicting cysteine S-Sulphenylation sites. Front Cell Dev Biol. 2020;8: 594587. https://doi.org/10.3389/fcell.2020.594587.

    Article  PubMed  PubMed Central  Google Scholar 

  47. Kawashima S, Ogata H, Kanehisa M. AAindex: amino acid index database. Nucleic Acids Res. 1999;27:368–9. https://doi.org/10.1093/nar/27.1.368.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  48. Henikoff S, Henikoff JG. Amino acid substitution matrices from protein blocks. Proc Natl Acad Sci U S A. 1992;89:10915–9. https://doi.org/10.1073/pnas.89.22.10915.

    Article  CAS  PubMed  PubMed Central  ADS  Google Scholar 

  49. Nair V, Hinton GE. Rectified linear units improve restricted boltzmann machines. In: Proceedings of the 27th international conference on machine learning (ICML-10). 2010, p. 807–814.

  50. Lin T-Y, Goyal P, Girshick R et al. Focal loss for dense object detection. In: Proceedings of the IEEE international conference on computer vision. 2017, p. 2980–2988.

  51. Kingma DP, Ba JJapa. Adam: A method for stochastic optimization 2014.

  52. Khanal J, Lim DY, Tayara H, et al. i6mA-stack: A stacking ensemble-based computational prediction of DNA N6-methyladenine (6mA) sites in the Rosaceae genome. Genomics. 2021;113:582–92. https://doi.org/10.1016/j.ygeno.2020.09.054.

    Article  CAS  PubMed  Google Scholar 

  53. Jia C, He W. EnhancerPred: a predictor for discovering enhancers based on the combination and selection of multiple features. Sci Rep. 2016;6:38741. https://doi.org/10.1038/srep38741.

    Article  CAS  PubMed  PubMed Central  ADS  Google Scholar 

  54. O’Shea JP, Chou MF, Quader SA, et al. pLogo: a probabilistic approach to visualizing sequence motifs. Nat Methods. 2013;10:1211–2. https://doi.org/10.1038/nmeth.2646.

    Article  CAS  PubMed  Google Scholar 

  55. Van der Maaten L, Hinton GJJomlr. Visualizing data using t-SNE 2008;9.

  56. Freund Y, Schapire REJJoc, sciences s. A decision-theoretic generalization of on-line learning and an application to boosting 1997;55:119-139.

Download references

Acknowledgements

We are very grateful to the experimental scientists for their publicly available data of Kcr sites.

Funding

This work was supported by the Start-up fund of Shanxi Normal University (83358).

Author information

Authors and Affiliations

Authors

Contributions

J.Y. devised the method, drafted and revised the manuscript. Y.R. analyzed the data and revised the manuscript. W.X. supervised the study and revised the manuscript.

Corresponding author

Correspondence to Xiaofeng Wang.

Ethics declarations

Ethics approval and consent to participate

Not applicable.

Consent for publication

Not applicable.

Competing interests

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary Information

Additional file 1.

Detailed information about the conventional machine learning models.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Jiang, Y., Yan, R. & Wang, X. PlantNh-Kcr: a deep learning model for predicting non-histone crotonylation sites in plants. Plant Methods 20, 28 (2024). https://doi.org/10.1186/s13007-024-01157-8

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s13007-024-01157-8

Keywords