Teja-Gollapudi commited on
Commit
b06b002
1 Parent(s): ea64011

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -4
README.md CHANGED
@@ -70,7 +70,10 @@ output = tokenizer.decode(output1[0])
70
 
71
  print(output)
72
 
73
- '''
 
 
 
74
  Sure, I can help you with that!
75
 
76
  Attention mechanisms in transformer models are typically implemented using the attention mechanism in the self-attention layer. Self-attention allows the model to focus on different parts of the input sequence when processing it. This is achieved by computing a set of attention weights, which are used to weigh the contribution of each input element to the output.
@@ -118,9 +121,8 @@ The `query`, `key`, and `value` tensors represent the input sequence to the tran
118
  The output of the `attention_weights` function is a NumPy tensor that represents the attention weights for the input sequence. These weights are used by the transformer model to weigh the contribution of each input element to the output.
119
 
120
  I hope this helps!</s>
121
- '''
122
- ```
123
-
124
  ## Finetuning details
125
  The finetuning scripts will be available in our [RAIL Github Repository](https://github.com/vmware-labs/research-and-development-artificial-intelligence-lab/tree/main/instruction-tuning)
126
  ## Evaluation
 
70
 
71
  print(output)
72
 
73
+ ```
74
+ ### Output
75
+
76
+
77
  Sure, I can help you with that!
78
 
79
  Attention mechanisms in transformer models are typically implemented using the attention mechanism in the self-attention layer. Self-attention allows the model to focus on different parts of the input sequence when processing it. This is achieved by computing a set of attention weights, which are used to weigh the contribution of each input element to the output.
 
121
  The output of the `attention_weights` function is a NumPy tensor that represents the attention weights for the input sequence. These weights are used by the transformer model to weigh the contribution of each input element to the output.
122
 
123
  I hope this helps!</s>
124
+ <hr>
125
+
 
126
  ## Finetuning details
127
  The finetuning scripts will be available in our [RAIL Github Repository](https://github.com/vmware-labs/research-and-development-artificial-intelligence-lab/tree/main/instruction-tuning)
128
  ## Evaluation