2023
DOI: 10.1093/bioinformatics/btad094
|View full text |Cite
|
Sign up to set email alerts
|

PFresGO: an attention mechanism-based deep-learning approach for protein annotation by integrating gene ontology inter-relationships

Abstract: Motivation The rapid accumulation of high-throughput sequence data demands the development of effective and efficient data-driven computational methods to functionally annotate proteins. However, most current approaches used for functional annotation simply focus on the use of protein-level information but ignore inter-relationships among annotations. Results Here, we established PFresGO, an attention-based deep-learning appr… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
7
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 16 publications
(10 citation statements)
references
References 29 publications
0
7
0
Order By: Relevance
“…Ontologies are not simply collections of classes; rather, ontologies are formal theories that specify some aspects of the intended meaning of a class using a logic-based language 23 . The background knowledge that is contained in the axioms of GO can be used by some machine learning models to improve predictions through knowledge-enhanced machine learning 11,12,14,15 . By incorporating the formal axioms into machine learning models, it becomes possible to leverage prior knowledge during the learning or prediction process, put constraints on the parameter search space that can improve the accuracy and efficiency of the learning process and, ultimately, make better predictions 24,25 .…”
Section: Uniprotkb/swiss-prot Dataset Evaluationmentioning
confidence: 99%
See 1 more Smart Citation
“…Ontologies are not simply collections of classes; rather, ontologies are formal theories that specify some aspects of the intended meaning of a class using a logic-based language 23 . The background knowledge that is contained in the axioms of GO can be used by some machine learning models to improve predictions through knowledge-enhanced machine learning 11,12,14,15 . By incorporating the formal axioms into machine learning models, it becomes possible to leverage prior knowledge during the learning or prediction process, put constraints on the parameter search space that can improve the accuracy and efficiency of the learning process and, ultimately, make better predictions 24,25 .…”
Section: Uniprotkb/swiss-prot Dataset Evaluationmentioning
confidence: 99%
“…These annotations are generally propagated to homologue proteins. As a result, the UniProtKB/Swiss-Prot database 3 contains manually curated GO annotations for thousands of organisms and more than 550,000 proteins.Recent protein function prediction methods rely on different sources of information such as sequence, interactions, protein tertiary structure, literature, coexpression, phylogenetic analysis or the information provided in GO [4][5][6][7][8][9][10][11][12][13][14][15][16][17][18][19][20] . The methods may use sequence domain annotations 5,6,8,11,21 , directly apply deep convolutional neural networks (CNN) 13 or language models such as long short-term memory neural networks 9 and transformers 14 , or use pretrained protein language models 10,15 to represent amino acid sequences.…”
mentioning
confidence: 99%
“…Recent protein function prediction methods rely on different sources of information such as sequence, interactions, protein tertiary structure, literature, coexpression, phylogenetic analysis, or the information provided in GO [4][5][6][7][8][9][10][11][12][13][14][15][16][17][18][19][20]. The methods may use sequence domain annotations [5,6,8,11,21], directly apply deep convolutional neural networks (CNN) [13] or language models such as LSTMs [9] and transformers [14], or use pretrained protein language models [10,15] to represent amino acid sequences. Models may also incorporate protein-protein interactions through knowledge graph embeddings [12,16], approaches using k-nearest neighbors [21], and graph convolutional neural networks [6].…”
Section: Introductionmentioning
confidence: 99%
“…Ontologies are not simply collections of classes; rather, ontologies are formal theories that specify some aspects of the intended meaning of a class using a logic-based language [23]. The background knowledge that is contained in the axioms of GO can be used by some machine learning models to improve predictions through knowledge-enhanced machine learning [11, 12, 14, 15]. By incorporating the formal axioms into machine learning models, it becomes possible to leverage prior knowledge during the learning or prediction process, and to put constraints on the parameter search space that can improve the accuracy and efficiency of the learning process, and, ultimately, make better predictions [24, 25].…”
Section: Introductionmentioning
confidence: 99%
“…Recent protein function prediction methods rely on different sources of information such as sequence, interactions, protein tertiary structure, literature, coexpression, phylogenetic analysis, or the information provided in GO [4][5][6][7][8][9][10][11][12][13][14][15][16][17][18][19][20]. The methods may use sequence domain annotations [5,6,8,11,21], directly apply deep convolutional neural networks (CNN) [13] or language models such as LSTMs [9] and transformers [14], or use pretrained protein language models [10,15] to represent amino acid sequences. Models may also incorporate protein-protein interactions through knowledge graph embeddings [12,16], approaches using k-nearest neighbors [21], and graph convolutional neural networks [6].…”
Section: Introductionmentioning
confidence: 99%