apcl
/

YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

Jam-Contextsum

Jam-Contextsum is a GPT2-like model finetuned to generate summary on why the method exists.

Jam-Contextsum Training Details

  • ckpt_pretrain is the file that we use to finetune the model for generating the summary on why the method exists
  • Our GitHub repo contains the code for reproduction using the same data.

ckpt_pretrain.pt

Hyperparameter Description Value
e embedding dimensions 512
L number of layers 4
h attention heads 4
c block size / context length 1,024
b batch size 4
a accumulation steps 32
d dropout 0.20
r learning rate 3e-5
y iterations 1e-5
iter number of iterations after pretraing 137,900
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.