Update README.md
Browse files
README.md
CHANGED
@@ -57,7 +57,7 @@ Assistant:
|
|
57 |
|
58 |
## How to use
|
59 |
|
60 |
-
### take the whole document as context
|
61 |
This can be applied to the scenario where the whole document can be fitted into the model, so that there is no need to run retrieval over the document.
|
62 |
```python
|
63 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
@@ -104,7 +104,7 @@ print(tokenizer.decode(response, skip_special_tokens=True))
|
|
104 |
```
|
105 |
|
106 |
### run retrieval to get top-n chunks as context
|
107 |
-
This can be applied to the scenario when the document is very long, so that it is necessary to run retrieval. Here, we use our [Dragon-multiturn](https://huggingface.co/nvidia/dragon-multiturn-query-encoder) retriever which can handle conversatinoal query. In addition, we provide a few [documents]() for users to play with.
|
108 |
|
109 |
```python
|
110 |
from transformers import AutoTokenizer, AutoModelForCausalLM, AutoModel
|
|
|
57 |
|
58 |
## How to use
|
59 |
|
60 |
+
### take the whole document as context
|
61 |
This can be applied to the scenario where the whole document can be fitted into the model, so that there is no need to run retrieval over the document.
|
62 |
```python
|
63 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
|
|
104 |
```
|
105 |
|
106 |
### run retrieval to get top-n chunks as context
|
107 |
+
This can be applied to the scenario when the document is very long, so that it is necessary to run retrieval. Here, we use our [Dragon-multiturn](https://huggingface.co/nvidia/dragon-multiturn-query-encoder) retriever which can handle conversatinoal query. In addition, we provide a few [documents](https://huggingface.co/nvidia/ChatQA-1.5-70B/tree/main/docs) for users to play with.
|
108 |
|
109 |
```python
|
110 |
from transformers import AutoTokenizer, AutoModelForCausalLM, AutoModel
|