kanashi6 commited on
Commit
01fb063
·
verified ·
1 Parent(s): 8f538aa

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +42 -0
README.md CHANGED
@@ -1,3 +1,45 @@
1
  ---
2
  license: apache-2.0
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: apache-2.0
3
+ language:
4
+ - en
5
  ---
6
+ # Model Card for Model ID
7
+
8
+ <!-- Provide a quick summary of what the model is/does. -->
9
+ [Beyond Language: Multi-layer Transformer is a General Visual Learner](https://arxiv.org/abs/2222.33333)
10
+ This repository includes ViM checkpoints, logs, and the pre-trained files used.
11
+
12
+ ## Model Details
13
+
14
+ ### Model Description
15
+
16
+ <!-- Provide a longer summary of what this model is. -->
17
+ In this project, we introduce ViM (Laqurge Visual Modeling). ViM has the following characteristics:
18
+
19
+ - 😮 **Minimalist architecture design similar to LLM**: ViM consists solely of a single transformer, without the inclusion of additional vision encoder and adapter.
20
+ - 🚀 **Covering all types of visual understanding tasks**: ViM addresses a spectrum of visual tasks, including object-level tasks (e.g., objecte detection), pixel-level tasks (e.g., semantic segmentation) and vision-language tasks (e.g., image captioning).
21
+ - 🤗 **Achieving task synergy by unified language interface**: Similar to LLM, ViM observes task synergy effect in multi-task training.
22
+ - 🔥 **SOTA performance on zero-shot and few-shot benchmark**: ViM scales well with model size and data, demonstrating remarkable generalizability across diverse scenarios after trained on 27 datasets.
23
+
24
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6585493b53c37507639fe3ba/FEhRT9ZscNwG7xIYYIYmh.png)
25
+
26
+ - **Developed by:** Haiyang Wang ( [email protected] ), Hao Tang ( [email protected] )
27
+ - **License:** [Apache license 2.0]
28
+
29
+ ### Model Sources [optional]
30
+
31
+ <!-- Provide the basic links for the model. -->
32
+
33
+ - **Repository:** https://github.com/Haiyang-W/ViM
34
+ - **Paper [optional]:** https://arxiv.org/abs/2222.33333
35
+
36
+ ## Uses
37
+ Please refer [here](https://github.com/Haiyang-W/ViM) for more detail about usage.
38
+
39
+
40
+
41
+
42
+
43
+
44
+
45
+