Teja-Gollapudi
commited on
Commit
•
b06b002
1
Parent(s):
ea64011
Update README.md
Browse files
README.md
CHANGED
@@ -70,7 +70,10 @@ output = tokenizer.decode(output1[0])
|
|
70 |
|
71 |
print(output)
|
72 |
|
73 |
-
|
|
|
|
|
|
|
74 |
Sure, I can help you with that!
|
75 |
|
76 |
Attention mechanisms in transformer models are typically implemented using the attention mechanism in the self-attention layer. Self-attention allows the model to focus on different parts of the input sequence when processing it. This is achieved by computing a set of attention weights, which are used to weigh the contribution of each input element to the output.
|
@@ -118,9 +121,8 @@ The `query`, `key`, and `value` tensors represent the input sequence to the tran
|
|
118 |
The output of the `attention_weights` function is a NumPy tensor that represents the attention weights for the input sequence. These weights are used by the transformer model to weigh the contribution of each input element to the output.
|
119 |
|
120 |
I hope this helps!</s>
|
121 |
-
|
122 |
-
|
123 |
-
|
124 |
## Finetuning details
|
125 |
The finetuning scripts will be available in our [RAIL Github Repository](https://github.com/vmware-labs/research-and-development-artificial-intelligence-lab/tree/main/instruction-tuning)
|
126 |
## Evaluation
|
|
|
70 |
|
71 |
print(output)
|
72 |
|
73 |
+
```
|
74 |
+
### Output
|
75 |
+
|
76 |
+
|
77 |
Sure, I can help you with that!
|
78 |
|
79 |
Attention mechanisms in transformer models are typically implemented using the attention mechanism in the self-attention layer. Self-attention allows the model to focus on different parts of the input sequence when processing it. This is achieved by computing a set of attention weights, which are used to weigh the contribution of each input element to the output.
|
|
|
121 |
The output of the `attention_weights` function is a NumPy tensor that represents the attention weights for the input sequence. These weights are used by the transformer model to weigh the contribution of each input element to the output.
|
122 |
|
123 |
I hope this helps!</s>
|
124 |
+
<hr>
|
125 |
+
|
|
|
126 |
## Finetuning details
|
127 |
The finetuning scripts will be available in our [RAIL Github Repository](https://github.com/vmware-labs/research-and-development-artificial-intelligence-lab/tree/main/instruction-tuning)
|
128 |
## Evaluation
|