<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Protein Structure on Personal Site</title><link>https://pdghawk.github.io/tags/protein-structure/</link><description>Recent content in Protein Structure on Personal Site</description><generator>Hugo -- gohugo.io</generator><language>en-us</language><lastBuildDate>Tue, 02 Jul 2024 11:55:56 -0400</lastBuildDate><atom:link href="https://pdghawk.github.io/tags/protein-structure/index.xml" rel="self" type="application/rss+xml"/><item><title>ESM3 is out: A first summary</title><link>https://pdghawk.github.io/posts/esm3/</link><pubDate>Tue, 02 Jul 2024 11:55:56 -0400</pubDate><guid>https://pdghawk.github.io/posts/esm3/</guid><description>On June 25th 2024 the ESM authors placed a preprint online (also on BioRxiv1 on July 2nd) titled &amp;ldquo;Simulating 500 million years of evolution with a language model&amp;rdquo;.
Key takeaways Using a masked language model, not an autoregressive model, with multiple tracks (sequence, structure) allows one to generate samples from any track Diffusion is not used for sampling structures, unlike Alphafold32. The tokenization, and generative procedure is performed via an attentive VQVAE model with iterative decoding scheme.</description></item></channel></rss>