Skip to content

IntelLabs/open-omics-alphafold

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

87 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

AlphaFold2 optimized on Intel Xeon CPU

Key words: Open-omics-alphaFold, AlphaFold2, AlphaFold2 on CPU, AlphaFold2 on Xeon, AlphaFold2 inference on SPR AVX512 FP32 and AMX-BF16

This repository contains an inference pipeline of AlphaFold2 with a bona fide translation from Haiku/JAX (https://github.com/deepmind/alphafold) to PyTorch.

Declaration 1 Any publication that discloses findings arising from using this source code or the model parameters should cite the AlphaFold paper. Please also refer to the Supplementary Information for a detailed description of the method.

Declaration 2 The setup procedures were modified from the two repos: https://github.com/kalininalab/alphafold_non_docker https://github.com/deepmind/alphafold with only some exceptions. I will label the difference for highlight.

Declaration 3 This repo is independently implemented, and is different from a previously unofficial version (https://github.com/lucidrains/alphafold2). No one is better than the other, and the differences are in 3 points: (1) this repo is major in acceleration of inference, in compatible to weights released from DeepMind; (2) this repo delivers a reliable pipeline accelerated on Intel® Core/Xeon and Intel® Optane® PMem by Intel® oneAPI. (3) this repo places CPU as its primary computation resource for acceleration, which may not provide an optimal speed on GPU.

Primary solution for setup of open-omics-alphafold environment

  1. Install miniforge;

      wget "https://github.com/conda-forge/miniforge/releases/latest/download/Miniforge3-$(uname)-$(uname -m).sh"
      bash Miniforge3-$(uname)-$(uname -m).sh
  2. create conda environment using a .yml file:

      conda env create -f conda_requirements.yml
      conda activate iaf2
  3. update submodules

      git submodule update --init --recursive
  4. Build dependencies for preprocessing (Optimized hh-suite and hmmer):

    build AVX512-optimized hh-suite

      export IAF2_DIR=`pwd`
      git clone --recursive https://github.com/IntelLabs/hh-suite.git
      cd hh-suite
      mkdir build && cd build
      cmake -DCMAKE_INSTALL_PREFIX=`pwd`/release -DCMAKE_CXX_COMPILER="icpx" -DCMAKE_CXX_FLAGS_RELEASE="-O3 -march=native" ..
      make -j 4 && make install
      ./release/bin/hhblits -h
      export PATH=`pwd`/release/bin:$PATH
      cd $IAF2_DIR

    build AVX512-optimized hmmer

      export IAF2_DIR=`pwd`
      git clone --recursive https://github.com/IntelLabs/hmmer.git
      cd hmmer
      cp easel_makefile.in easel/Makefile.in
      cd easel && make clean && autoconf && ./configure --prefix=`pwd` && cd ..
      autoconf && CC=icx CFLAGS="-O3 -march=native -fPIC" ./configure --prefix=`pwd`/release
      make -j 4 && make install
      ./release/bin/jackhmmer -h
      export PATH=`pwd`/release/bin:$PATH
      cd $IAF2_DIR
  5. Install jemalloc

      git clone --branch 5.3.0 https://github.com/jemalloc/jemalloc.git
      cd jemalloc && bash autogen.sh --prefix=$CONDA_PREFIX && make install 
      cd ..
      export LD_LIBRARY_PATH="$CONDA_PREFIX/lib:$LD_LIBRARY_PATH"
  6. build dependency for TPP optimization of AlphaFold2 [Global]Attention Modules:

    TPP-pytorch-extension implements efficient kernels for Xeon CPUs in C++ using the libxsmm library. If setup failed, AlphaFold2 will fall back to enable PyTorch JIT w/o PCL-extension. (Use gcc > 11.4.0)

    export IAF2_DIR=`pwd`
    git clone https://github.com/libxsmm/tpp-pytorch-extension
    cd tpp-pytorch-extension
    git submodule update --init
    CC=gcc CXX=g++ python setup.py install
    python -c "from tpp_pytorch_extension.alphafold.Alpha_Attention import GatingAttentionOpti_forward"
  7. extract weights in the <root_home> directory

    mkdir weights && mkdir weights/extracted
    python extract_params.py --input <data-dir>/params/params_model_1.npz --output_dir ./weights/extracted/model_1

    For Multimer,

    python extract_params.py --input <data-dir>/params/model_1_multimer_v3.npz --output_dir ./weights/extracted/model_1_multimer_v3
  8. Put your query sequence files in "<input-dir>" folder:

    all fasta sequences should be named as *.fa 1 sequence per each file, e.g. example.fa

     > example file
     ATGCCGCATGGTCGTC
    

    For Multimer, remember to put each chain separately in the fasta file, even when the chains are identical. The following file shows two identical chains in a multimer fasta file.

    >6E3K_1|Chains A|Interferon gamma|Homo sapiens (9606)
     GPGSQDPYVKEAENLKKYFNAGHSDVADNGTLFLGILKNWKEESDRKIMQSQIVSFYFKLFKNFKDDQSIQKSVETIKEDMNVKFFNSNKKKRDDFEKLTNYSVTDLNVQRKAIHELIQVMAELSPAAKTGKRKRSQAAAHHHHHHHH
     >6E3K_1|Chains B|Interferon gamma|Homo sapiens (9606)
     GPGSQDPYVKEAENLKKYFNAGHSDVADNGTLFLGILKNWKEESDRKIMQSQIVSFYFKLFKNFKDDQSIQKSVETIKEDMNVKFFNSNKKKRDDFEKLTNYSVTDLNVQRKAIHELIQVMAELSPAAKTGKRKRSQAAAHHHHHHHH
    
  9. run main scripts to test your env

    run preprocess main script to do MSA and template search on 1st sample in $root_home/samples

      bash online_preproc_monomer.sh <root_home> <data-dir> <input-dir> <output-dir>
      # please ensure your query sequence files *.fa are in <input-dir>

    For Multimer,

    bash online_preproc_multimer.sh <root_home> <data-dir> <input-dir> <output-dir>

    intermediates data can be seen under $output-dir//intermediates and $output-dir//msa

    run model inference script to predict unrelaxed structures from MSA and template results

    bash online_inference_monomer.sh <conda_env_path> <root_home> <input-dir> <output-dir> <model_names>

    For Multimer,

    bash online_inference_multimer.sh <conda_env_path> <root_home> <input-dir> <output-dir> <model_names>

    By default, inference runs in bfloat16 precision. Set AF2_BF16 input to 0 to run in FP32 precision. unrelaxed data can be seen under $output-dir/

  10. Run relaxation script (Untested)

    Download stereo_chemical_props.txt file into alphafold/common folder using the following command

    wget -q -P ./alphafold/common/ https://git.scicore.unibas.ch/schwede/openstructure/-/raw/7102c63615b64735c4941278d92b554ec94415f8/modules/mol/alg/src/stereo_chemical_props.txt --no-check-certificate

    Run the relaxation script with the following command

    bash one_amber.sh <root_home> <input-dir> <output-dir> <model_names> monomer

    For Multimer,

    bash one_amber.sh <root_home> <input-dir> <output-dir> <model_names> multimer <num_multimer_predictions_per_model>
  11. Multi-instance Throughput Run First, create a logs directory in the <root_home> directory with the following command

    mkdir <root_home>/logs

    Run the multi-instance preprocessing script with the following command

    python run_multiprocess_pre.py --root_home=<root_home> --data_dir=<data_dir> --input_dir=<input_dir> --output_dir=<output_dir> --model_name=<model_name>

    For Multimer,

    python run_multiprocess_pre_multimer.py --root_home=<root_home> --data_dir=<data_dir> --input_dir=<input_dir> --output_dir=<output_dir>

    Set library paths correctly export LD_LIBRARY_PATH=$CONDA_PREFIX/lib:$LD_LIBRARY_PATH export LD_PRELOAD=$CONDA_PREFIX/lib/libjemalloc.so:$LD_PRELOAD Run the multi-instance model inference script with the following command

    python run_multiprocess_infer.py --root_home=<root_home> --input_dir=<input_dir> --output_dir=<output_dir> --model_names=<model_names>

    For Multimer,

    python run_multiprocess_infer_multimer.py --root_home=<root_home> --input_dir=<input_dir> --output_dir=<output_dir> --model_names=<model_names>

    For multiprocess relaxation,

    python run_multiprocess_relax.py --root_home=<root_home> --input_dir=<input_dir> --output_dir=<output_dir> --model_names=<model_names> --model_preset=monomer   # or multimer

All steps are ended here for optimized AlphaFold2.

The following lines are stock information of the Original Alphafold2 repo:

  1. Update is on schedule: AlphaFold with Multimers will be coming soon

Genetic databases

Please use the AlphaFold2 database download instructions from the original repo. (https://github.com/google-deepmind/alphafold?tab=readme-ov-file#genetic-databases)

Model parameters

While the AlphaFold code is licensed under the Apache 2.0 License, the AlphaFold parameters are made available for non-commercial use only under the terms of the CC BY-NC 4.0 license. Please see the Disclaimer below for more detail.

Please follow instructions from the original AlphaFold2 repo. https://github.com/google-deepmind/alphafold?tab=readme-ov-file#model-parameters

Running AlphaFold

Recommended server configuration

  1. CPU: 2-sockets, Intel® Xeon® Scalable Performance Processor (61xx, 81xx, 62xx, 82xx, 92xx, 63xx, 83xx, etc.)
  2. Memory: DRAM >192GB, or Intel® Optane® Persistent Memory (PMem) for higher Memory (e.g. 6TB/socket)
  3. Disk: Intel® Optane® SSD

We need to extract the original model parameters into directory tree, so that PyTorch version of Alphafold2 can easily load params w/o mistakes. Please use extract_params.py to execute such convertion.

AlphaFold output

The outputs will be in a subfolder of output_dir in run_docker.py. They include the computed MSAs, unrelaxed structures, relaxed structures, ranked structures, raw model outputs, prediction metadata, and section timings. The output_dir directory will have the following structure:

<target_name>/
    features.pkl
    ranked_{0,1,2,3,4}.pdb
    ranking_debug.json
    relaxed_model_{1,2,3,4,5}.pdb
    result_model_{1,2,3,4,5}.pkl
    timings.json
    unrelaxed_model_{1,2,3,4,5}.pdb
    msas/
        bfd_uniclust_hits.a3m
        mgnify_hits.sto
        uniref90_hits.sto
    intermediates/
        features.npz
        processed_features.npz

The contents of each output file are as follows:

  • features.pkl – A pickle file containing the input feature NumPy arrays used by the models to produce the structures.

  • unrelaxed_model_*.pdb – A PDB format text file containing the predicted structure, exactly as outputted by the model.

  • relaxed_model_*.pdb – A PDB format text file containing the predicted structure, after performing an Amber relaxation procedure on the unrelaxed structure prediction (see Jumper et al. 2021, Suppl. Methods 1.8.6 for details).

  • ranked_*.pdb – A PDB format text file containing the relaxed predicted structures, after reordering by model confidence. Here ranked_0.pdb should contain the prediction with the highest confidence, and ranked_4.pdb the prediction with the lowest confidence. To rank model confidence, we use predicted LDDT (pLDDT) scores (see Jumper et al. 2021, Suppl. Methods 1.9.6 for details).

  • ranking_debug.json – A JSON format text file containing the pLDDT values used to perform the model ranking, and a mapping back to the original model names.

  • timings.json – A JSON format text file containing the times taken to run each section of the AlphaFold pipeline.

  • msas/ - A directory containing the files describing the various genetic tool hits that were used to construct the input MSA.

  • result_model_*.pkl – A pickle file containing a nested dictionary of the various NumPy arrays directly produced by the model. In addition to the output of the structure module, this includes auxiliary outputs such as:

    • Distograms (distogram/logits contains a NumPy array of shape [N_res, N_res, N_bins] and distogram/bin_edges contains the definition of the bins).
    • Per-residue pLDDT scores (plddt contains a NumPy array of shape [N_res] with the range of possible values from 0 to 100, where 100 means most confident). This can serve to identify sequence regions predicted with high confidence or as an overall per-target confidence score when averaged across residues.
    • Present only if using pTM models: predicted TM-score (ptm field contains a scalar). As a predictor of a global superposition metric, this score is designed to also assess whether the model is confident in the overall domain packing.
    • Present only if using pTM models: predicted pairwise aligned errors (predicted_aligned_error contains a NumPy array of shape [N_res, N_res] with the range of possible values from 0 to max_predicted_aligned_error, where 0 means most confident). This can serve for a visualisation of domain packing confidence within the structure.

The pLDDT confidence measure is stored in the B-factor field of the output PDB files (although unlike a B-factor, higher pLDDT is better, so care must be taken when using for tasks such as molecular replacement).

This code has been tested to match mean top-1 accuracy on a CASP14 test set with pLDDT ranking over 5 model predictions (some CASP targets were run with earlier versions of AlphaFold and some had manual interventions; see our forthcoming publication for details). Some targets such as T1064 may also have high individual run variance over random seeds.

Citing this work

If you use the code or data in this package, please cite:

@Article{AlphaFold2021,
  author  = {Jumper, John and Evans, Richard and Pritzel, Alexander and Green, Tim and Figurnov, Michael and Ronneberger, Olaf and Tunyasuvunakool, Kathryn and Bates, Russ and {\v{Z}}{\'\i}dek, Augustin and Potapenko, Anna and Bridgland, Alex and Meyer, Clemens and Kohl, Simon A A and Ballard, Andrew J and Cowie, Andrew and Romera-Paredes, Bernardino and Nikolov, Stanislav and Jain, Rishub and Adler, Jonas and Back, Trevor and Petersen, Stig and Reiman, David and Clancy, Ellen and Zielinski, Michal and Steinegger, Martin and Pacholska, Michalina and Berghammer, Tamas and Bodenstein, Sebastian and Silver, David and Vinyals, Oriol and Senior, Andrew W and Kavukcuoglu, Koray and Kohli, Pushmeet and Hassabis, Demis},
  journal = {Nature},
  title   = {Highly accurate protein structure prediction with {AlphaFold}},
  year    = {2021},
  doi     = {10.1038/s41586-021-03819-2},
  note    = {(Accelerated article preview)},
}

Community contributions

Colab notebooks provided by the community (please note that these notebooks may vary from our full AlphaFold system and we did not validate their accuracy):

Acknowledgements

AlphaFold communicates with and/or references the following separate libraries and packages:

We thank all their contributors and maintainers!

License and Disclaimer

This is not an officially supported Google product.

Copyright 2021 DeepMind Technologies Limited.

AlphaFold Code License

Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at https://www.apache.org/licenses/LICENSE-2.0.

Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License.

Model Parameters License

The AlphaFold parameters are made available for non-commercial use only, under the terms of the Creative Commons Attribution-NonCommercial 4.0 International (CC BY-NC 4.0) license. You can find details at: https://creativecommons.org/licenses/by-nc/4.0/legalcode

Third-party software

Use of the third-party software, libraries or code referred to in the Acknowledgements section above may be governed by separate terms and conditions or license provisions. Your use of the third-party software, libraries or code is subject to any such terms and you should check that you can comply with any applicable restrictions or terms and conditions before use.

Mirrored Databases

The following databases have been mirrored by DeepMind, and are available with reference to the following: