jglaser's picture
initial commit
1d7245f

How to use the data sets

This dataset contains more about 80,000 unique pairs of protein sequences and ligand SMILES, and the coordinates of their complexes from the PDB. Only ligands with a molecular weight >= 100 Da are included.

SMILES are assumed to be tokenized by the regex from P. Schwaller.

Every (x,y,z) ligand coordinate maps onto a SMILES token, and is nan if the token does not represent an atom

Every receptor coordinate maps onto the Calpha coordinate of that residue.

The dataset can be used to fine-tune a language model, all data comes from PDBind-cn.

Use the already preprocessed data

Load a test/train split using

from datasets import load_dataset
train = load_dataset("jglaser/pdb_protein_ligand_complexes",split='train[:90%]')
validation = load_dataset("jglaser/pdb_protein_ligand_complexes",split='train[90%:]')

Manual update from PDB

# download the PDB archive into folder pdb/
sh rsync.sh 24 # number of parallel download processes

# extract sequences and coordinates in parallel
sbatch pdb.slurm
# or
mpirun -n 42 parse_complexes.py # desired number of tasks