Skip to preprint detailsSkip to PREreviews

PREreviews of InterPLM: Discovering Interpretable Features in Protein Language Models via Sparse Autoencoders

1 PREreview

  1. PREreview by Karson Chrispens et al.

    Summary

    This study seeks to apply the recently developed techniques for mechanistic interpretability of large language models to protein language models, namely ESM-2. By training a sparse autoencoder (SAE) to reconstruct embeddings from the trunk of ESM-2-8M, the authors claim to extract…

    Read the PREreview by Karson Chrispens et al.