Files changed (1) hide show
  1. README.md +8 -7
README.md CHANGED
@@ -30,14 +30,14 @@ REXzyme (Reaction to Enzyme) (manuscript in preparation) is a translation machin
30
  It is possible to provide fine-grained input at the substrate level.
31
  Akin to how translation machines have learned to translate between complex language pairs with great success,
32
  often diverging in their representation at the character level, (Japanese - English), we posit that an advanced architecture will
33
- be able to translate between the chemical and sequence spaces. REXzyme was trained on a set of xx reactions and yy enzyme pairs and it produces
34
  sequences that putatitely perform their intended reactions.
35
 
36
  To run it, you will need to provide a reaction in the SMILE format (Simplified molecular-input line-entry system), which you can do online here: https://cactus.nci.nih.gov/chemical/structure.
37
 
38
- After converting each of the reaction components you should combine them in the following scheme : ReactantA.ReactantB>AgentA>ProductA.ProductB
39
- Additionally prepend the task suffix "r2s" and append the eos token "</s>"
40
- e.g. for the carbonic anhydrase "r2sO.COO>>HCOOO.[H+]</s>"
41
 
42
  or via this simple python script:
43
 
@@ -75,9 +75,9 @@ and contains 48 (24 el/ 24 dl) layers with a model dimensionality of 1024, total
75
 
76
  REXzyme is a translation machine trained on portion the RHEA database containing 31970152 reaction-enzyme pairs.
77
  The pre-training was done on pairs of smiles and amino acid sequences, tokenized with a char-level
78
- Sentencepiece tokenizer. Note that two seperate tokenizers were used for input and labels.
79
 
80
- REXzyme was pre-trained with a supervised translation objective i.e., the model learned to use the continous representation of the reaction from the encoder to autoregressivly (causual language modeling) produce the output trying to match the shifted right target enzyme sequence. Hence, the model learns the dependencies among protein sequence features that enable a specific enzymatic reaction.
81
 
82
  There are stark differences in the number of members among Reaction classes, and for this reason. Since we are tokenizing the reaction smiles on a char level, classes with few reactions can profit from the knwodledge gained for classes catalyzing similar reactions that have a lot of members.
83
 
@@ -87,6 +87,7 @@ The figure below summarizes the process of training: (add figure) [STILL MISSING
87
  ## **Model Performance**
88
 
89
  - **Dataset curation**
 
90
  <br/><br/>
91
  - **General descriptors**
92
 
@@ -94,7 +95,7 @@ The figure below summarizes the process of training: (add figure) [STILL MISSING
94
  | :--- | :----: | ---: |
95
  | **IUPRED3 (ordered)** | 99.9% | 99.9% |
96
  | **ESMFold** | 85.03 | 71.59 (selected: 79.82) |
97
- | **FlDPnn** | missing | missing |
98
  | **PSIpred** | missing | missing |
99
  <br/><br/>
100
 
 
30
  It is possible to provide fine-grained input at the substrate level.
31
  Akin to how translation machines have learned to translate between complex language pairs with great success,
32
  often diverging in their representation at the character level, (Japanese - English), we posit that an advanced architecture will
33
+ be able to translate between the chemical and sequence spaces. REXzyme was trained on a set of 2480 reactions and ~32M enzyme pairs and it produces
34
  sequences that putatitely perform their intended reactions.
35
 
36
  To run it, you will need to provide a reaction in the SMILE format (Simplified molecular-input line-entry system), which you can do online here: https://cactus.nci.nih.gov/chemical/structure.
37
 
38
+ After converting each of the reaction components you should combine them in the following scheme : ```ReactantA.ReactantB>AgentA>ProductA.ProductB```<br/>
39
+ Additionally prepend the task suffix ```r2s``` and append the eos token ```</s>```
40
+ e.g. for the carbonic anhydrase ```r2sO.COO>>HCOOO.[H+]</s>```
41
 
42
  or via this simple python script:
43
 
 
75
 
76
  REXzyme is a translation machine trained on portion the RHEA database containing 31970152 reaction-enzyme pairs.
77
  The pre-training was done on pairs of smiles and amino acid sequences, tokenized with a char-level
78
+ Sentencepiece tokenizer. Note that two seperate tokenizers were used for input (smiles_tokenizer) and labels (aa_tokenizer).
79
 
80
+ REXzyme was pre-trained with a supervised translation objective i.e., the model learned to use the continous representation of the reaction from the encoder to autoregressivly (causual language modeling) produce the output by shifting it right on token (amino acid) at a time trying to match the target enzyme sequence. Hence, the model learns the dependencies among protein sequence features that enable a specific enzymatic reaction.
81
 
82
  There are stark differences in the number of members among Reaction classes, and for this reason. Since we are tokenizing the reaction smiles on a char level, classes with few reactions can profit from the knwodledge gained for classes catalyzing similar reactions that have a lot of members.
83
 
 
87
  ## **Model Performance**
88
 
89
  - **Dataset curation**
90
+ We converted the reactions from rxn format to smile string including only left-to-right reactions. The enzyme sequences were truncated to 1024. Enzymes catalyzing more than one reaction were given multiple enzyme-reaction entries.
91
  <br/><br/>
92
  - **General descriptors**
93
 
 
95
  | :--- | :----: | ---: |
96
  | **IUPRED3 (ordered)** | 99.9% | 99.9% |
97
  | **ESMFold** | 85.03 | 71.59 (selected: 79.82) |
98
+ | **FlDPnn** | missing | 0.0929 |
99
  | **PSIpred** | missing | missing |
100
  <br/><br/>
101