2019
DOI: 10.1038/s41586-019-1335-8
|View full text |Cite
|
Sign up to set email alerts
|

Unsupervised word embeddings capture latent knowledge from materials science literature

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

4
689
0
11

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 835 publications
(704 citation statements)
references
References 38 publications
4
689
0
11
Order By: Relevance
“…Furthermore, the vast majority of databases are commercial products requiring a license, and programmatic application programming interfaces (APIs) for large‐scale data access are rarely implemented. A large fraction of experimental data are only available in journal publications, though recent successes in text mining offer a potential solution to this conundrum . Finally, major efforts are underway in high‐throughput/combinatorial experiments that can generate large experimental materials database with diverse properties …”
Section: Data Collectionmentioning
confidence: 99%
See 4 more Smart Citations
“…Furthermore, the vast majority of databases are commercial products requiring a license, and programmatic application programming interfaces (APIs) for large‐scale data access are rarely implemented. A large fraction of experimental data are only available in journal publications, though recent successes in text mining offer a potential solution to this conundrum . Finally, major efforts are underway in high‐throughput/combinatorial experiments that can generate large experimental materials database with diverse properties …”
Section: Data Collectionmentioning
confidence: 99%
“…While PCA works on linear projection, the manifold learning is able to capture nonlinear relationships. For example, the t‐distributed stochastic neighbor embedding (t‐SNE) method learns low‐dimensional representations such that the local distance between data points is roughly preserved and has been applied in visualizing the elemental embedding vector trained from materials property prediction models, structural similarity of perovskites, word embeddings in text mining, electronic fingerprints, etc.…”
Section: Featurizationmentioning
confidence: 99%
See 3 more Smart Citations