TransPTM : A transformer-based model for non-histone acetylation site prediction
Research output: Journal Publications and Reviews › RGC 21 - Publication in refereed journal › peer-review
Author(s)
Related Research Unit(s)
Detail(s)
Original language | English |
---|---|
Article number | bbae219 |
Number of pages | 10 |
Journal / Publication | Briefings in Bioinformatics |
Volume | 25 |
Issue number | 3 |
Publication status | Published - May 2024 |
Link(s)
DOI | DOI |
---|---|
Attachment(s) | Documents
Publisher's Copyright Statement
|
Link to Scopus | https://www.scopus.com/record/display.uri?eid=2-s2.0-85192792843&origin=recordpage |
Permanent Link | https://scholars.cityu.edu.hk/en/publications/publication(119eebf1-730b-473a-8a3b-9056b99c66d7).html |
Abstract
Protein acetylation is one of the extensively studied post-translational modifications (PTMs) due to its significant roles across a myriad of biological processes. Although many computational tools for acetylation site identification have been developed, there is a lack of benchmark dataset and bespoke predictors for non-histone acetylation site prediction.To address these problems,we have contributed to both dataset creation and predictor benchmark in this study. First, we construct a non-histone acetylation site benchmark dataset, namely NHAC, which includes 11 subsets according to the sequence length ranging from 11 to 61 amino acids. There are totally 886 positive samples and 4707 negative samples for each sequence length. Secondly, we propose TransPTM, a transformer-based neural network model for non-histone acetylation site predication. During the data representation phase, per-residue contextualized embeddings are extracted using ProtT5 (an existing pre-trained protein language model). This is followed by the implementation of a graph neural network framework, which consists of three TransformerConv layers for feature extraction and a multilayer perceptron module for classification. The benchmark results reflect that TransPTM has the competitive performance for non-histone acetylation site prediction over three state-of-the-art tools.It improves our comprehension on the PTM mechanism and provides a theoretical basis for developing drug targets for diseases. Moreover, the created PTM datasets fills the gap in non-histone acetylation site datasets and is beneficial to the related communities. The related source code and data utilized by TransPTM are accessible at https://www.github. com/TransPTM/TransPTM.
© The Author(s) 2024. Published by Oxford University Press.
© The Author(s) 2024. Published by Oxford University Press.
Research Area(s)
- deep learning, Non-histone acetylation, protein language model, transformer
Citation Format(s)
TransPTM: A transformer-based model for non-histone acetylation site prediction. / Meng, Lingkuan; Chen, Xingjian; Cheng, Ke et al.
In: Briefings in Bioinformatics, Vol. 25, No. 3, bbae219, 05.2024.
In: Briefings in Bioinformatics, Vol. 25, No. 3, bbae219, 05.2024.
Research output: Journal Publications and Reviews › RGC 21 - Publication in refereed journal › peer-review
Download Statistics
No data available