zihanliu commited on
Commit
725e93f
·
verified ·
1 Parent(s): 4ab852d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -57,7 +57,7 @@ Assistant:
57
 
58
  ## How to use
59
 
60
- ### take the whole document as context
61
  This can be applied to the scenario where the whole document can be fitted into the model, so that there is no need to run retrieval over the document.
62
  ```python
63
  from transformers import AutoTokenizer, AutoModelForCausalLM
@@ -104,7 +104,7 @@ print(tokenizer.decode(response, skip_special_tokens=True))
104
  ```
105
 
106
  ### run retrieval to get top-n chunks as context
107
- This can be applied to the scenario when the document is very long, so that it is necessary to run retrieval. Here, we use our [Dragon-multiturn](https://huggingface.co/nvidia/dragon-multiturn-query-encoder) retriever which can handle conversatinoal query. In addition, we provide a few [documents]() for users to play with.
108
 
109
  ```python
110
  from transformers import AutoTokenizer, AutoModelForCausalLM, AutoModel
 
57
 
58
  ## How to use
59
 
60
+ ### take the whole document as context
61
  This can be applied to the scenario where the whole document can be fitted into the model, so that there is no need to run retrieval over the document.
62
  ```python
63
  from transformers import AutoTokenizer, AutoModelForCausalLM
 
104
  ```
105
 
106
  ### run retrieval to get top-n chunks as context
107
+ This can be applied to the scenario when the document is very long, so that it is necessary to run retrieval. Here, we use our [Dragon-multiturn](https://huggingface.co/nvidia/dragon-multiturn-query-encoder) retriever which can handle conversatinoal query. In addition, we provide a few [documents](https://huggingface.co/nvidia/ChatQA-1.5-70B/tree/main/docs) for users to play with.
108
 
109
  ```python
110
  from transformers import AutoTokenizer, AutoModelForCausalLM, AutoModel