“…Bepler and Berger ( 70 ) pretrained LSTMs on protein sequences, adding supervision from contacts to produce embeddings. Subsequent to our preprint, related works have built on its exploration of protein sequence modeling, exploring generative models ( 71 , 72 ), internal representations of Transformers ( 73 ), and applications of representation learning and generative modeling such as classification ( 74 , 75 ), mutational effect prediction ( 80 ), and design of sequences ( 76 ā 78 ).…”