chunwoolee0 commited on
Commit
2792eff
1 Parent(s): 21b535a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -6
README.md CHANGED
@@ -8,12 +8,11 @@ model-index:
8
  results: []
9
  ---
10
 
11
- <!-- This model card has been generated automatically according to the information Keras had access to. You should
12
- probably proofread and complete it, then remove this comment. -->
13
 
14
  # chunwoolee0/distilroberta-base-finetuned-wikitext2
15
 
16
- This model is a fine-tuned version of [distilroberta-base](https://huggingface.co/distilroberta-base) on an unknown dataset.
 
17
  It achieves the following results on the evaluation set:
18
  - Train Loss: 2.1557
19
  - Validation Loss: 1.8964
@@ -21,18 +20,19 @@ It achieves the following results on the evaluation set:
21
 
22
  ## Model description
23
 
24
- More information needed
25
 
26
  ## Intended uses & limitations
27
 
28
- More information needed
29
 
30
  ## Training and evaluation data
31
 
32
- More information needed
33
 
34
  ## Training procedure
35
 
 
36
  ### Training hyperparameters
37
 
38
  The following hyperparameters were used during training:
 
8
  results: []
9
  ---
10
 
 
 
11
 
12
  # chunwoolee0/distilroberta-base-finetuned-wikitext2
13
 
14
+ This model is a fine-tuned version of [distilroberta-base](https://huggingface.co/distilroberta-base)
15
+ on an [wikitext,wikitext-2-raw-v1](https://huggingface.co/datasets/wikitext/viewer/wikitext-2-raw-v1/test) dataset.
16
  It achieves the following results on the evaluation set:
17
  - Train Loss: 2.1557
18
  - Validation Loss: 1.8964
 
20
 
21
  ## Model description
22
 
23
+ This model is a distilled version of the RoBERTa-base model. It follows the same training procedure as DistilBERT.
24
 
25
  ## Intended uses & limitations
26
 
27
+ This is an exercise for finetuning of nlp language modeling for fill-mask.
28
 
29
  ## Training and evaluation data
30
 
31
+ Wikitext, wikitext-2-raw-v1 is used
32
 
33
  ## Training procedure
34
 
35
+
36
  ### Training hyperparameters
37
 
38
  The following hyperparameters were used during training: