Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

how is alpha fold 3 trained #1031

Open
joepareti54 opened this issue Oct 16, 2024 · 0 comments
Open

how is alpha fold 3 trained #1031

joepareti54 opened this issue Oct 16, 2024 · 0 comments

Comments

@joepareti54
Copy link

joepareti54 commented Oct 16, 2024

This paper says that the loss function is the weighted sum of L_distogram, L_diffusion and L_confidence. But how is it implemented? taking the derivative to update the weights ? When the processing is in the trunk there is no diffusion yet which occurs later in time. When the processing is in diffusion, the trunk does not seem to execute again. Is there any backpropagation? Moreover the diffusion process training involves predicting the noise injected at a time step, but then the only loss term to base the weight updates ought to be Ldiffusion. How about the attention layers and MLP training ?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant