FrameFlow is a SE(3) flow matching method for protein backbone generation and motif-scaffolding. The method is described in two papers:
- Improved motif-scaffolding with SE(3) flow matching
- Fast protein backbone generation with SE (3) flow matching
Unconditional protein backbone generation:
Motif condifioned scaffold backbone generation:
If you use this work (or code), then please cite the first paper (citing both would make me happier :).
@article{
yim2024improved,
title={Improved motif-scaffolding with {SE}(3) flow matching},
author={Jason Yim and Andrew Campbell and Emile Mathieu and Andrew Y. K. Foong and Michael Gastegger and Jose Jimenez-Luna and Sarah Lewis and Victor Garcia Satorras and Bastiaan S. Veeling and Frank Noe and Regina Barzilay and Tommi Jaakkola},
journal={Transactions on Machine Learning Research},
issn={2835-8856},
year={2024},
url={https://openreview.net/forum?id=fa1ne8xDGn},
note={}
}
@article{yim2023fast,
title={Fast protein backbone generation with SE (3) flow matching},
author={Yim, Jason and Campbell, Andrew and Foong, Andrew YK and Gastegger, Michael and Jim{\'e}nez-Luna, Jos{\'e} and Lewis, Sarah and Satorras, Victor Garcia and Veeling, Bastiaan S and Barzilay, Regina and Jaakkola, Tommi and others},
journal={arXiv preprint arXiv:2310.05297},
year={2023}
}
Note
See the legacy branch for the old version of this repo from the workshop paper Fast protein backbone generation with SE (3) flow matching.
Adding ProteinMPNN or folding with AlphaFold2 or ESMFold can't be done currently with Microsoft's open source policy for this project. I am in discussion for a more standardized motif-scaffolding codebase specifically for benchmarking that myself (and others) hope will make this task easier to evaluate. Currently it is very arduous with requiring over 2600 ProteinMPNN and AlphaFold2 evaluations for each benchmark run.
Set up your conda environment.
# Conda environment with dependencies.
conda env create -f fm.yml
# Activate environment
conda activate fm
# Manually need to install torch-scatter.
pip install torch-scatter -f https://data.pyg.org/whl/torch-2.0.0+cu117.html
# Install local package.
# Current directory should be protein-frame-flow/
pip install -e .
Datasets and weights are hosted on Zenodo here.
preprocessed_pdb.tar.gz
(2.7 GB)weights.tar.gz
(0.6 GB)preprocessed_scope.tar.gz
(0.3 GB)
Next, untar the datasets
tar -xzvf preprocessed_pdb.tar.gz
tar -xzvf weights.tar.gz
tar -xzvf preprocessed_scope.tar.gz
Other datasets are also possible to train on by processing with the data/process_pdb_files.py
script.
Your directory should now look like this
├── analysis
├── build
├── configs
├── data
├── experiments
├── media
├── models
├── openfold
├── processed_pdb
├── processed_scope
└── weights
Our training relies on logging with wandb. Log in to Wandb and make an account. Authorize Wandb here.
All training flags are in configs/base.yaml
. Below is explanation-by-example of the main flags to change. Note you can combine multiple flags in the command.
# Train on PDB or SCOPE
python -W ignore experiments/train_se3_flows.py data.dataset=pdb
python -W ignore experiments/train_se3_flows.py data.dataset=scope
# Train on hallucination or inpainting (motif-scaffolding)
python -W ignore experiments/train_se3_flows.py data.task=hallucination
python -W ignore experiments/train_se3_flows.py data.inpainting=inpainting
# Train with larger batches. Depends on GPU memory.
python -W ignore experiments/train_se3_flows.py data.sampler.max_num_res_squared=600_000
# Train with more GPUs
python -W ignore experiments/train_se3_flows.py experiment.num_devices=4
Our inference script allows for DDP. By default we sample 10 sequences per length [70, 100, 200, 300]. Samples are stored as PDB files as well as the trajectories. We do not include evaluation code using ProteinMPNN and ESMFold but this should be easy to set-up if one looks at the FrameDiff codebase.
# Single GPU
python -W ignore experiments/inference_se3_flows.py -cn inference_unconditional
# Multi GPU
python -W ignore experiments/inference_se3_flows.py -cn inference_unconditional inference.num_gpus=2
By default, we use and recommend the amortization model.
Like unconditional sampling, multi-GPU DDP is supported.
We support the RFdiffusion motif-scaffolding benchmark as described in Supp. Methods Table 9 of Watson et al.
We do not include motif 6VW1 since it involves multiple chains that FrameFlow cannot handle.
Benchmarking works by reading the contig settings in motif_scaffolding/benchmark.csv
.
Consider target 3IXT
in the CSV.
target | contig | length | motif_path |
---|---|---|---|
3IXT | 10-40,P254-277,10-40 | 50-75 | ./motif_scaffolding/targets/3IXT.pdb |
Explanation of each column:
target
: unique identifier for the motif.contig
: same contig syntax as RFdiffusion to specify motifs from a PDB file and scaffolds to sample. See the motif-scaffolding section in the RFdiffusion README for more information.length
: randomly sampled total scaffold length.motif_path
: path to PDB file with motif.
Note
To specify your own motif, follow the syntax in benchmark.csv
and point samples.csv_path
to the your custom CSV with motif-scaffolding tasks.
To run motif-scaffolding, we specify the settings in configs/inference_scaffolding.yaml
.
See inference.samples
for different sampling settings.
# Single GPU
python -W ignore experiments/inference_se3_flows.py -cn inference_scaffolding
# Multi GPU
python -W ignore experiments/inference_se3_flows.py -cn inference_scaffolding inference.num_gpus=2
# FrameFlow-guidance
python -W ignore experiments/inference_se3_flows.py -cn inference_scaffolding_guidance
- What is FrameFlow?
- FrameFlow is a deep neural network that models 3D protein structures.
- What can FrameFlow do?
- By sampling from FrameFlow, you can obtain a description of the positions and orientations of the backbone atoms in a protein.
- What is/are FrameFlow’s intended use(s)?
- FrameFlow is intended for research purposes only, for the machine learning for structural biology community.
- How was FrameFlow evaluated? What metrics are used to measure performance?
- FrameFlow was evaluated on how novel, designable and diverse the protein structures sampled from FrameFlow were.
- What are the limitations of FrameFlow? How can users minimize the impact of FrameFlow’s limitations when using the system?
- FrameFlow has not been tested by real-world experiments to see if the proteins it samples are actually designable. FrameFlow should be used for research purposes only.
- What operational factors and settings allow for effective and responsible use of FrameFlow?
- FrameFlow should be used for research purposes only.
This project welcomes contributions and suggestions. Most contributions require you to agree to a Contributor License Agreement (CLA) declaring that you have the right to, and actually do, grant us the rights to use your contribution. For details, visit https://cla.opensource.microsoft.com.
When you submit a pull request, a CLA bot will automatically determine whether you need to provide a CLA and decorate the PR appropriately (e.g., status check, comment). Simply follow the instructions provided by the bot. You will only need to do this once across all repos using our CLA.
This project has adopted the Microsoft Open Source Code of Conduct. For more information see the Code of Conduct FAQ or contact [email protected] with any additional questions or comments.
This project may contain trademarks or logos for projects, products, or services. Authorized use of Microsoft trademarks or logos is subject to and must follow Microsoft's Trademark & Brand Guidelines. Use of Microsoft trademarks or logos in modified versions of this project must not cause confusion or imply Microsoft sponsorship. Any use of third-party trademarks or logos are subject to those third-party's policies.