Daniele Licari commited on
Commit
56a273a
·
1 Parent(s): 69d53a2

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -18,7 +18,7 @@ Archive using the Huggingface PyTorch-Transformers library. We used BERT archite
18
  with a language modeling head on top, AdamW Optimizer, initial learning rate 5e-5 (with
19
  linear learning rate decay, ends at 2.525e-9), sequence length 512, batch size 10 (imposed
20
  by GPU capacity), 8.4 million training steps, device 1*GPU V100 16GB
21
-
22
  <h2> Usage </h2>
23
 
24
  ITALIAN-LEGAL-BERT model can be loaded like:
 
18
  with a language modeling head on top, AdamW Optimizer, initial learning rate 5e-5 (with
19
  linear learning rate decay, ends at 2.525e-9), sequence length 512, batch size 10 (imposed
20
  by GPU capacity), 8.4 million training steps, device 1*GPU V100 16GB
21
+ <p />
22
  <h2> Usage </h2>
23
 
24
  ITALIAN-LEGAL-BERT model can be loaded like: