doberst commited on
Commit
bd3de1e
1 Parent(s): 8d7c5ba

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -48
README.md CHANGED
@@ -33,56 +33,10 @@ Note: compare results with [dragon-mistral-7b](https://www.huggingface.co/llmwar
33
  <!-- Provide a longer summary of what this model is. -->
34
 
35
  - **Developed by:** llmware
36
- - **Model type:** bling-rag-instruct
37
  - **Language(s) (NLP):** English
38
  - **License:** Apache 2.0
39
- - **Finetuned from model:** Microsoft Phi-3
40
-
41
- ## Uses
42
-
43
- <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
44
-
45
- The intended use of BLING models is two-fold:
46
-
47
- 1. Provide high-quality RAG-Instruct models designed for fact-based, no "hallucination" question-answering in connection with an enterprise RAG workflow.
48
-
49
- 2. BLING models are fine-tuned on top of leading base foundation models, generally in the 1-3B+ range, and purposefully rolled-out across multiple base models to provide choices and "drop-in" replacements for RAG specific use cases.
50
-
51
-
52
- ### Direct Use
53
-
54
- <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
55
-
56
- BLING is designed for enterprise automation use cases, especially in knowledge-intensive industries, such as financial services,
57
- legal and regulatory industries with complex information sources.
58
-
59
- BLING models have been trained for common RAG scenarios, specifically: question-answering, key-value extraction, and basic summarization as the core instruction types
60
- without the need for a lot of complex instruction verbiage - provide a text passage context, ask questions, and get clear fact-based responses.
61
-
62
-
63
- ## Bias, Risks, and Limitations
64
-
65
- <!-- This section is meant to convey both technical and sociotechnical limitations. -->
66
-
67
- BLING models are designed to operate with grounded sources, e.g., inclusion of a context passage in the prompt, and will not yield consistent or positive results if open-context prompting in which you are looking for the model to draw upon potential background knowledge of the world - in fact, it is likely that the BLING will respond with a simple "Not Found." to an open context query.
68
-
69
- Any model can provide inaccurate or incomplete information, and should be used in conjunction with appropriate safeguards and fact-checking mechanisms.
70
-
71
-
72
- ## How to Get Started with the Model
73
-
74
- To pull the model via API:
75
-
76
- from huggingface_hub import snapshot_download
77
- snapshot_download("llmware/bling-phi-3-gguf", local_dir="/path/on/your/machine/", local_dir_use_symlinks=False)
78
-
79
- Load in your favorite GGUF inference engine, or try with llmware as follows:
80
-
81
- from llmware.models import ModelCatalog
82
-
83
- # to load the model and make a basic inference
84
- model = ModelCatalog().load_model("llmware/bling-phi-3-gguf", temperature=0.0, sample=False)
85
- response = model.inference(query, add_context=text_sample)
86
 
87
  Details on the prompt wrapper and other configurations are on the config.json file in the files repository.
88
 
 
33
  <!-- Provide a longer summary of what this model is. -->
34
 
35
  - **Developed by:** llmware
36
+ - **Model type:** dragon-rag-instruct
37
  - **Language(s) (NLP):** English
38
  - **License:** Apache 2.0
39
+ - **Finetuned from model:** Mistral-7B-0.3-Base
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
40
 
41
  Details on the prompt wrapper and other configurations are on the config.json file in the files repository.
42