muhtasham's picture
Librarian Bot: Add base_model information to model (#2)
fe6a78c
|
raw
history blame
690 Bytes
metadata
datasets:
  - albertvillanova/legal_contracts
base_model: google/bert_uncased_L-2_H-128_A-2

bert-tiny-finetuned-legal-contracts-longer

This model is a fine-tuned version of google/bert_uncased_L-2_H-128_A-2 on the portion of legal_contracts dataset but for longer epochs.

Note

The model was not trained on the whole dataset which is around 9.5 GB, but only

The first 10% of train + the last 10% of train.

datasets_train = load_dataset('albertvillanova/legal_contracts' , split='train[:10%]')
datasets_validation = load_dataset('albertvillanova/legal_contracts' , split='train[-10%:]')