Upload README.md
Browse files
README.md
CHANGED
@@ -8,7 +8,7 @@ language:
|
|
8 |
|
9 |
## Model Details
|
10 |
|
11 |
-
model-card-testing is a distilled language
|
12 |
|
13 |
- **Developed by:** author1, author2
|
14 |
- **Model type:** testing type
|
@@ -22,7 +22,15 @@ model-card-testing is a distilled language model. Users of this model card shoul
|
|
22 |
|
23 |
## How to Get Started with the Model
|
24 |
|
25 |
-
Use the code below to get started with the model.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
26 |
|
27 |
|
28 |
|
@@ -52,16 +60,16 @@ encoded_input = tokenizer(text, return_tensors='tf')
|
|
52 |
output = model(encoded_input)
|
53 |
```
|
54 |
|
55 |
-
##
|
56 |
LLMs are intended to be used for language generation or as a pretrained base model that can be further fine-tuned for specific tasks. Use cases below are not exhaustive.
|
57 |
|
58 |
-
|
59 |
This model can be used for:
|
60 |
- Text generation
|
61 |
- Exploring characterisitics of language generated by a language model
|
62 |
- Examples: Cloze tests, counterfactuals, generations with reframings
|
63 |
|
64 |
-
|
65 |
Tasks that leverage language models, including:
|
66 |
- Information Extraction, Question Answering, Summarization
|
67 |
- Writing assistance: Grammar assistance, autocompletion (for normal prose or code)
|
@@ -71,7 +79,7 @@ Tasks that leverage language models, including:
|
|
71 |
### Misuse and Out-of-scope Use
|
72 |
Using the model in high-stakes settings is out of scope for this model. The model is not designed for critical decisions nor uses with any material consequences on an individual's livelihood or wellbeing. The model outputs content that appears factual but is not correct.
|
73 |
|
74 |
-
|
75 |
|
76 |
**CONTENT WARNING: Readers should be aware this section contains content that is disturbing, offensive, and can propogate historical and current stereotypes.**
|
77 |
|
@@ -89,14 +97,26 @@ The impact of model compression techniques, such as knowledge distillation, on b
|
|
89 |
NOTE: This code will need customization/fixing.
|
90 |
|
91 |
|
|
|
|
|
|
|
|
|
|
|
92 |
|
|
|
|
|
|
|
93 |
|
94 |
|
95 |
-
|
|
|
|
|
|
|
|
|
96 |
|
97 |
model-card-testing was trained using . See the data card for additional information.
|
98 |
|
99 |
-
|
100 |
|
101 |
Preprocessing, hardware used, hyperparameters...
|
102 |
|
@@ -107,6 +127,11 @@ This model achieves the following results:
|
|
107 |
NOTE: This will need customization.
|
108 |
|
109 |
|
|
|
|
|
|
|
|
|
|
|
110 |
|
111 |
|
112 |
|
@@ -120,7 +145,7 @@ You can estimate carbon emissions using the [Machine Learning Impact calculator]
|
|
120 |
- **Compute Region:**
|
121 |
- **Carbon Emitted** *(Power consumption x Time x Carbon produced based on location of power grid)*:
|
122 |
|
123 |
-
|
124 |
|
125 |
```bibtex
|
126 |
@inproceedings{...,
|
|
|
8 |
|
9 |
## Model Details
|
10 |
|
11 |
+
model-card-testing is a distilled language modelthat can be used for text generation. Users of this model card should also consider information about the design, training, and limitations of gpt2.
|
12 |
|
13 |
- **Developed by:** author1, author2
|
14 |
- **Model type:** testing type
|
|
|
22 |
|
23 |
## How to Get Started with the Model
|
24 |
|
25 |
+
Use the code below to get started with the model. model-card-testing can be used directly with a pipeline for text generation.
|
26 |
+
Since the generation relies on some randomness, we set a seed for reproducibility:
|
27 |
+
```python
|
28 |
+
>>> from transformers import pipeline, set_seed
|
29 |
+
>>> generator = pipeline('text-generation', model='model-card-testing')
|
30 |
+
>>> set_seed(42)
|
31 |
+
>>> generator("Hello, I'm a language model," max_length=20, num_return_sequences=5)
|
32 |
+
```
|
33 |
+
|
34 |
|
35 |
|
36 |
|
|
|
60 |
output = model(encoded_input)
|
61 |
```
|
62 |
|
63 |
+
## Uses
|
64 |
LLMs are intended to be used for language generation or as a pretrained base model that can be further fine-tuned for specific tasks. Use cases below are not exhaustive.
|
65 |
|
66 |
+
### Direct Use
|
67 |
This model can be used for:
|
68 |
- Text generation
|
69 |
- Exploring characterisitics of language generated by a language model
|
70 |
- Examples: Cloze tests, counterfactuals, generations with reframings
|
71 |
|
72 |
+
### Downstream Use
|
73 |
Tasks that leverage language models, including:
|
74 |
- Information Extraction, Question Answering, Summarization
|
75 |
- Writing assistance: Grammar assistance, autocompletion (for normal prose or code)
|
|
|
79 |
### Misuse and Out-of-scope Use
|
80 |
Using the model in high-stakes settings is out of scope for this model. The model is not designed for critical decisions nor uses with any material consequences on an individual's livelihood or wellbeing. The model outputs content that appears factual but is not correct.
|
81 |
|
82 |
+
## Limitations
|
83 |
|
84 |
**CONTENT WARNING: Readers should be aware this section contains content that is disturbing, offensive, and can propogate historical and current stereotypes.**
|
85 |
|
|
|
97 |
NOTE: This code will need customization/fixing.
|
98 |
|
99 |
|
100 |
+
```python
|
101 |
+
>>> from transformers import pipeline, set_seed
|
102 |
+
>>> generator = pipeline('text-generation', model='model-card-testing')
|
103 |
+
>>> set_seed(48)
|
104 |
+
>>> generator("The White man worked as a", max_length=20, num_return_sequences=3)
|
105 |
|
106 |
+
>>> set_seed(48)
|
107 |
+
>>> generator("The Black man worked as a", max_length=20, num_return_sequences=3)
|
108 |
+
```
|
109 |
|
110 |
|
111 |
+
|
112 |
+
|
113 |
+
## Training
|
114 |
+
|
115 |
+
### Training Data
|
116 |
|
117 |
model-card-testing was trained using . See the data card for additional information.
|
118 |
|
119 |
+
### Training Procedure
|
120 |
|
121 |
Preprocessing, hardware used, hyperparameters...
|
122 |
|
|
|
127 |
NOTE: This will need customization.
|
128 |
|
129 |
|
130 |
+
| Dataset | LAMBADA | LAMBADA | CBT-CN | CBT-NE | WikiText2 | PTB | enwiki8 | text8 | WikiText103 | 1BW |
|
131 |
+
|:--------:|:-------:|:-------:|:------:|:------:|:---------:|:------:|:-------:|:------:|:-----------:|:-----:|
|
132 |
+
| (metric) | (PPL) | (ACC) | (ACC) | (ACC) | (PPL) | (PPL) | (BPB) | (BPC) | (PPL) | (PPL) |
|
133 |
+
| | | | | | | | | | | |
|
134 |
+
|
135 |
|
136 |
|
137 |
|
|
|
145 |
- **Compute Region:**
|
146 |
- **Carbon Emitted** *(Power consumption x Time x Carbon produced based on location of power grid)*:
|
147 |
|
148 |
+
## Citation Information
|
149 |
|
150 |
```bibtex
|
151 |
@inproceedings{...,
|