|
--- |
|
license: cc-by-4.0 |
|
datasets: |
|
- bltlab/queryner |
|
language: |
|
- en |
|
metrics: |
|
- f1 |
|
pipeline_tag: token-classification |
|
inference: |
|
parameters: |
|
aggregation_strategy: "first" |
|
--- |
|
|
|
# Model Card for Model ID |
|
|
|
E-commerce query segmentation model in English. |
|
This model is trained on QueryNER training dataset with the addition of augmentations so the model should be more robust to spelling mistakes and mentions unseen in the training data. |
|
|
|
|
|
## Model Details |
|
|
|
### Model Description |
|
|
|
This is a token classification model using BERT base uncased as the base model. |
|
The model is fine-tuned on the (QueryNER training dataset)[https://huggingface.co/datasets/bltlab/queryner] and augmented data as described in the QueryNER paper. |
|
|
|
|
|
- **Developed by:** [BLT Lab](https://github.com/bltlab) in collaboration with eBay. |
|
- **Funded by:** eBay |
|
- **Shared by:** (@cpalenmichel)[https://github.com/cpalenmichel] |
|
- **Model type:** Token Classification / Sequence Labeling / Chunking |
|
- **Language(s) (NLP):** English |
|
- **License:** CC-BY 4.0 |
|
- **Finetuned from model:** BERT base uncased |
|
|
|
### Model Sources |
|
|
|
Underlying model is based on [BERT base-uncased](https://huggingface.co/google-bert/bert-base-uncased). |
|
|
|
- **Repository:** [https://github.com/bltlab/query-ner](https://github.com/bltlab/query-ner) |
|
- **Paper:** Accepted at LREC-COLING Coming soon |
|
|
|
## Uses |
|
|
|
### Direct Use |
|
|
|
Intended use is research purposes and e-commerce query segmentation. |
|
|
|
### Downstream Use |
|
|
|
Potential downstream use cases include weighting entity spans, linking to knowledge bases, removing spans as a recovery strategy for null and low recall queries. |
|
|
|
### Out-of-Scope Use |
|
|
|
This model is trained only on the training data of the QueryNER dataset. It may not perform well on other domains without additional training data and further fine-tuning. |
|
|
|
## Bias, Risks, and Limitations |
|
|
|
See paper limitations section. |
|
|
|
## How to Get Started with the Model |
|
|
|
See huggingface tutorials for token classification and access the model using AutoModelForTokenClassification. |
|
Note that we do some post processing to make use of only the first subtoken's tag unlike the inference API. |
|
|
|
## Training Details |
|
|
|
### Training Data |
|
|
|
See paper for details. |
|
|
|
|
|
### Training Procedure |
|
|
|
See paper for details. |
|
|
|
#### Training Hyperparameters |
|
|
|
See paper for details. |
|
|
|
|
|
## Evaluation |
|
|
|
Evaluation details provided in the paper. |
|
Scoring was done using [SeqScore](https://github.com/bltlab/seqscore) using the conlleval repair method for invalid label transition sequences. |
|
|
|
### Testing Data, Factors & Metrics |
|
|
|
#### Testing Data |
|
|
|
QueryNER test set: [https://huggingface.co/datasets/bltlab/queryner](https://huggingface.co/datasets/bltlab/queryner) |
|
|
|
|
|
#### Factors |
|
Evaluation is reported with micro-F1 at the entity level on the QueryNER test set. |
|
We used conlleval repair method for invalid label transitions. |
|
|
|
#### Metrics |
|
We use micro-F1 at the entity level as this is fairly common practice for NER models. |
|
|
|
### Results |
|
|
|
[More Information Needed] |
|
|
|
|
|
## Environmental Impact |
|
Rough estimate |
|
|
|
- **Hardware Type:** 1 RTX 3090 GPU |
|
- **Hours used:** < 2 hours |
|
- **Cloud Provider:** Private |
|
- **Compute Region:** northamerica-northeast1 |
|
- **Carbon Emitted:** 0.02 |
|
|
|
|
|
## Citation |
|
|
|
Accepted at LREC-COLING coming soon |
|
|
|
**BibTeX:** |
|
|
|
Accepted at LREC-COLING coming soon |
|
|
|
|
|
## Model Card Authors |
|
|
|
Chester Palen-Michel (@cpalenmichel)[https://github.com/cpalenmichel] |
|
|
|
## Model Card Contact |
|
|
|
Chester Palen-Michel (@cpalenmichel)[https://github.com/cpalenmichel] |