Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
@@ -50,14 +50,14 @@ def query_image(img, text_queries, score_threshold):
|
|
50 |
|
51 |
|
52 |
description = """
|
53 |
-
|
54 |
-
introduced in <a href="https://arxiv.org/abs/
|
55 |
-
|
56 |
-
|
57 |
To use it, simply upload an image and enter comma separated text descriptions of objects you want to query the image for. You
|
58 |
can also use the score threshold slider to set a threshold to filter out low probability predictions.
|
59 |
\n\nOWL-ViT is trained on text templates,
|
60 |
-
hence you can get better predictions by querying the image with text templates used in training the original model: *"photo of a star-spangled banner"*,
|
61 |
*"image of a shoe"*. Refer to the <a href="https://arxiv.org/abs/2103.00020">CLIP</a> paper to see the full list of text templates used to augment the training data.
|
62 |
\n\n<a href="https://colab.research.google.com/github/huggingface/notebooks/blob/main/examples/zeroshot_object_detection_with_owlvit.ipynb">Colab demo</a>
|
63 |
"""
|
|
|
50 |
|
51 |
|
52 |
description = """
|
53 |
+
Try this demo for <a href="https://huggingface.co/docs/transformers/main/en/model_doc/owlv2">OWLv2</a>,
|
54 |
+
introduced in <a href="https://arxiv.org/abs/2306.09683">Scaling Open-Vocabulary Object Detection</a>.
|
55 |
+
\n\n Compared to OWLVIT, OWLv2 performs better both in yield and performance (average precision).
|
56 |
+
You can use OWLv2 to query images with text descriptions of any object.
|
57 |
To use it, simply upload an image and enter comma separated text descriptions of objects you want to query the image for. You
|
58 |
can also use the score threshold slider to set a threshold to filter out low probability predictions.
|
59 |
\n\nOWL-ViT is trained on text templates,
|
60 |
+
hence you can get better predictions by querying the image with text templates used in training the original model: e.g. *"photo of a star-spangled banner"*,
|
61 |
*"image of a shoe"*. Refer to the <a href="https://arxiv.org/abs/2103.00020">CLIP</a> paper to see the full list of text templates used to augment the training data.
|
62 |
\n\n<a href="https://colab.research.google.com/github/huggingface/notebooks/blob/main/examples/zeroshot_object_detection_with_owlvit.ipynb">Colab demo</a>
|
63 |
"""
|