Papers
arxiv:2401.02987

Has Your Pretrained Model Improved? A Multi-head Posterior Based Approach

Published on Jan 2, 2024
· Submitted by akhaliq on Jan 9, 2024
Authors:
,
,
,
,
,

Abstract

The emergence of pretrained models has significantly impacted from Natural Language Processing (NLP) and Computer Vision to <PRE_TAG>relational datasets</POST_TAG>. Traditionally, these models are assessed through fine-tuned downstream tasks. However, this raises the question of how to evaluate these models more efficiently and more effectively. In this study, we explore a novel approach where we leverage the meta features associated with each entity as a source of worldly knowledge and employ entity representations from the models. We propose using the consistency between these representations and the meta features as a metric for evaluating pretrained models. Our method's effectiveness is demonstrated across various domains, including models with <PRE_TAG>relational datasets</POST_TAG>, large language models and images models.

Community

Sign up or log in to comment

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2401.02987 in a model README.md to link it from this page.

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2401.02987 in a dataset README.md to link it from this page.

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2401.02987 in a Space README.md to link it from this page.

Collections including this paper 0

No Collection including this paper

Add this paper to a collection to link it from this page.