Update README.md
Browse files
README.md
CHANGED
@@ -17,7 +17,7 @@ LLaVA is an open-source chatbot trained by fine-tuning LLaMA/Vicuna on GPT-gener
|
|
17 |
It is an auto-regressive language model, based on the transformer architecture.
|
18 |
|
19 |
**Model date:**
|
20 |
-
LLaVA-
|
21 |
|
22 |
**Paper or resources for more information:**
|
23 |
https://llava-vl.github.io/
|
@@ -38,12 +38,12 @@ The primary intended users of the model are researchers and hobbyists in compute
|
|
38 |
|
39 |
## Training dataset
|
40 |
- 558K filtered image-text pairs from LAION/CC/SBU, captioned by BLIP.
|
41 |
-
-
|
|
|
|
|
42 |
|
43 |
## Evaluation dataset
|
44 |
-
A
|
45 |
-
We also evaluate our model on the ScienceQA dataset. Our synergy with GPT-4 sets a new state-of-the-art on the dataset.
|
46 |
-
See https://llava-vl.github.io/ for more details.
|
47 |
|
48 |
## Usage
|
49 |
usage is as follows
|
|
|
17 |
It is an auto-regressive language model, based on the transformer architecture.
|
18 |
|
19 |
**Model date:**
|
20 |
+
LLaVA-v1.5-7B was trained in September 2023.
|
21 |
|
22 |
**Paper or resources for more information:**
|
23 |
https://llava-vl.github.io/
|
|
|
38 |
|
39 |
## Training dataset
|
40 |
- 558K filtered image-text pairs from LAION/CC/SBU, captioned by BLIP.
|
41 |
+
- 158K GPT-generated multimodal instruction-following data.
|
42 |
+
- 450K academic-task-oriented VQA data mixture.
|
43 |
+
- 40K ShareGPT data.
|
44 |
|
45 |
## Evaluation dataset
|
46 |
+
A collection of 12 benchmarks, including 5 academic VQA benchmarks and 7 recent benchmarks specifically proposed for instruction-following LMMs.
|
|
|
|
|
47 |
|
48 |
## Usage
|
49 |
usage is as follows
|