## How to use the data sets This dataset contains more about 80,000 unique pairs of protein sequences and ligand SMILES, and the coordinates of their complexes from the PDB. Only ligands with a molecular weight >= 100 Da are included. SMILES are assumed to be tokenized by the regex from P. Schwaller. Every (x,y,z) ligand coordinate maps onto a SMILES token, and is *nan* if the token does not represent an atom Every receptor coordinate maps onto the Calpha coordinate of that residue. The dataset can be used to fine-tune a language model, all data comes from PDBind-cn. ### Use the already preprocessed data Load a test/train split using ``` from datasets import load_dataset train = load_dataset("jglaser/pdb_protein_ligand_complexes",split='train[:90%]') validation = load_dataset("jglaser/pdb_protein_ligand_complexes",split='train[90%:]') ``` ### Manual update from PDB ``` # download the PDB archive into folder pdb/ sh rsync.sh 24 # number of parallel download processes # extract sequences and coordinates in parallel sbatch pdb.slurm # or mpirun -n 42 parse_complexes.py # desired number of tasks ```