jysh1023 commited on
Commit
5bbc4ab
Β·
1 Parent(s): 8be076d

Training in progress, epoch 1

Browse files
logs/events.out.tfevents.1700313495.0a848c80699a.8533.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0789b7d98277a9ee5d21fac6763b32e22a4a43aff0183db36f09da5dfcd2e06e
3
+ size 4715
nncf_output.log CHANGED
@@ -45,53 +45,6 @@ INFO:nncf:Collecting tensor statistics |β–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 1 / 3
45
  INFO:nncf:Collecting tensor statistics |β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 2 / 3
46
  INFO:nncf:Collecting tensor statistics |β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 3 / 3
47
  INFO:nncf:Compiling and loading torch extension: quantized_functions_cuda...
48
- INFO:nncf:Not adding activation input quantizer for operation: 7 BertForSequenceClassification/BertModel[bert]/BertEmbeddings[embeddings]/NNCFEmbedding[position_embeddings]/embedding_0
49
- INFO:nncf:Not adding activation input quantizer for operation: 4 BertForSequenceClassification/BertModel[bert]/BertEmbeddings[embeddings]/NNCFEmbedding[word_embeddings]/embedding_0
50
- INFO:nncf:Not adding activation input quantizer for operation: 5 BertForSequenceClassification/BertModel[bert]/BertEmbeddings[embeddings]/NNCFEmbedding[token_type_embeddings]/embedding_0
51
- INFO:nncf:Not adding activation input quantizer for operation: 6 BertForSequenceClassification/BertModel[bert]/BertEmbeddings[embeddings]/__add___0
52
- INFO:nncf:Not adding activation input quantizer for operation: 8 BertForSequenceClassification/BertModel[bert]/BertEmbeddings[embeddings]/__iadd___0
53
- INFO:nncf:Not adding activation input quantizer for operation: 9 BertForSequenceClassification/BertModel[bert]/BertEmbeddings[embeddings]/NNCFLayerNorm[LayerNorm]/layer_norm_0
54
- INFO:nncf:Not adding activation input quantizer for operation: 10 BertForSequenceClassification/BertModel[bert]/BertEmbeddings[embeddings]/Dropout[dropout]/dropout_0
55
- INFO:nncf:Not adding activation input quantizer for operation: 23 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[0]/BertAttention[attention]/BertSelfAttention[self]/__add___0
56
- INFO:nncf:Not adding activation input quantizer for operation: 26 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[0]/BertAttention[attention]/BertSelfAttention[self]/matmul_1
57
- INFO:nncf:Not adding activation input quantizer for operation: 32 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[0]/BertAttention[attention]/BertSelfOutput[output]/__add___0
58
- INFO:nncf:Not adding activation input quantizer for operation: 33 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[0]/BertAttention[attention]/BertSelfOutput[output]/NNCFLayerNorm[LayerNorm]/layer_norm_0
59
- INFO:nncf:Not adding activation input quantizer for operation: 37 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[0]/BertOutput[output]/__add___0
60
- INFO:nncf:Not adding activation input quantizer for operation: 38 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[0]/BertOutput[output]/NNCFLayerNorm[LayerNorm]/layer_norm_0
61
- INFO:nncf:Not adding activation input quantizer for operation: 51 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[1]/BertAttention[attention]/BertSelfAttention[self]/__add___0
62
- INFO:nncf:Not adding activation input quantizer for operation: 54 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[1]/BertAttention[attention]/BertSelfAttention[self]/matmul_1
63
- INFO:nncf:Not adding activation input quantizer for operation: 60 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[1]/BertAttention[attention]/BertSelfOutput[output]/__add___0
64
- INFO:nncf:Not adding activation input quantizer for operation: 61 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[1]/BertAttention[attention]/BertSelfOutput[output]/NNCFLayerNorm[LayerNorm]/layer_norm_0
65
- INFO:nncf:Not adding activation input quantizer for operation: 65 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[1]/BertOutput[output]/__add___0
66
- INFO:nncf:Not adding activation input quantizer for operation: 66 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[1]/BertOutput[output]/NNCFLayerNorm[LayerNorm]/layer_norm_0
67
- INFO:nncf:Not adding activation input quantizer for operation: 79 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[2]/BertAttention[attention]/BertSelfAttention[self]/__add___0
68
- INFO:nncf:Not adding activation input quantizer for operation: 82 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[2]/BertAttention[attention]/BertSelfAttention[self]/matmul_1
69
- INFO:nncf:Not adding activation input quantizer for operation: 88 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[2]/BertAttention[attention]/BertSelfOutput[output]/__add___0
70
- INFO:nncf:Not adding activation input quantizer for operation: 89 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[2]/BertAttention[attention]/BertSelfOutput[output]/NNCFLayerNorm[LayerNorm]/layer_norm_0
71
- INFO:nncf:Not adding activation input quantizer for operation: 93 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[2]/BertOutput[output]/__add___0
72
- INFO:nncf:Not adding activation input quantizer for operation: 94 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[2]/BertOutput[output]/NNCFLayerNorm[LayerNorm]/layer_norm_0
73
- INFO:nncf:Not adding activation input quantizer for operation: 107 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[3]/BertAttention[attention]/BertSelfAttention[self]/__add___0
74
- INFO:nncf:Not adding activation input quantizer for operation: 110 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[3]/BertAttention[attention]/BertSelfAttention[self]/matmul_1
75
- INFO:nncf:Not adding activation input quantizer for operation: 116 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[3]/BertAttention[attention]/BertSelfOutput[output]/__add___0
76
- INFO:nncf:Not adding activation input quantizer for operation: 117 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[3]/BertAttention[attention]/BertSelfOutput[output]/NNCFLayerNorm[LayerNorm]/layer_norm_0
77
- INFO:nncf:Not adding activation input quantizer for operation: 121 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[3]/BertOutput[output]/__add___0
78
- INFO:nncf:Not adding activation input quantizer for operation: 122 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[3]/BertOutput[output]/NNCFLayerNorm[LayerNorm]/layer_norm_0
79
- INFO:nncf:Not adding activation input quantizer for operation: 135 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[4]/BertAttention[attention]/BertSelfAttention[self]/__add___0
80
- INFO:nncf:Not adding activation input quantizer for operation: 138 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[4]/BertAttention[attention]/BertSelfAttention[self]/matmul_1
81
- INFO:nncf:Not adding activation input quantizer for operation: 144 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[4]/BertAttention[attention]/BertSelfOutput[output]/__add___0
82
- INFO:nncf:Not adding activation input quantizer for operation: 145 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[4]/BertAttention[attention]/BertSelfOutput[output]/NNCFLayerNorm[LayerNorm]/layer_norm_0
83
- INFO:nncf:Not adding activation input quantizer for operation: 149 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[4]/BertOutput[output]/__add___0
84
- INFO:nncf:Not adding activation input quantizer for operation: 150 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[4]/BertOutput[output]/NNCFLayerNorm[LayerNorm]/layer_norm_0
85
- INFO:nncf:Not adding activation input quantizer for operation: 163 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[5]/BertAttention[attention]/BertSelfAttention[self]/__add___0
86
- INFO:nncf:Not adding activation input quantizer for operation: 166 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[5]/BertAttention[attention]/BertSelfAttention[self]/matmul_1
87
- INFO:nncf:Not adding activation input quantizer for operation: 172 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[5]/BertAttention[attention]/BertSelfOutput[output]/__add___0
88
- INFO:nncf:Not adding activation input quantizer for operation: 173 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[5]/BertAttention[attention]/BertSelfOutput[output]/NNCFLayerNorm[LayerNorm]/layer_norm_0
89
- INFO:nncf:Not adding activation input quantizer for operation: 177 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[5]/BertOutput[output]/__add___0
90
- INFO:nncf:Not adding activation input quantizer for operation: 178 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[5]/BertOutput[output]/NNCFLayerNorm[LayerNorm]/layer_norm_0
91
- INFO:nncf:Collecting tensor statistics |β–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 1 / 3
92
- INFO:nncf:Collecting tensor statistics |β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 2 / 3
93
- INFO:nncf:Collecting tensor statistics |β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 3 / 3
94
- INFO:nncf:Compiling and loading torch extension: quantized_functions_cuda...
95
  INFO:nncf:Finished loading torch extension: quantized_functions_cuda
96
  WARNING:nncf:You are setting `forward` on an NNCF-processed model object.
97
  NNCF relies on custom-wrapping the `forward` call in order to function properly.
@@ -170,19 +123,48 @@ model.nncf.set_original_unbound_forward(fn)
170
  if `fn` has an unbound 0-th `self` argument, or
171
  with model.nncf.temporary_bound_original_forward(fn): ...
172
  if `fn` already had 0-th `self` argument bound or never had it in the first place.
173
- WARNING:nncf:You are setting `forward` on an NNCF-processed model object.
174
- NNCF relies on custom-wrapping the `forward` call in order to function properly.
175
- Arbitrary adjustments to the forward function on an NNCFNetwork object have undefined behavior.
176
- If you need to replace the underlying forward function of the original model so that NNCF should be using that instead of the original forward function that NNCF saved during the compressed model creation, you can do this by calling:
177
- model.nncf.set_original_unbound_forward(fn)
178
- if `fn` has an unbound 0-th `self` argument, or
179
- with model.nncf.temporary_bound_original_forward(fn): ...
180
- if `fn` already had 0-th `self` argument bound or never had it in the first place.
181
- WARNING:nncf:You are setting `forward` on an NNCF-processed model object.
182
- NNCF relies on custom-wrapping the `forward` call in order to function properly.
183
- Arbitrary adjustments to the forward function on an NNCFNetwork object have undefined behavior.
184
- If you need to replace the underlying forward function of the original model so that NNCF should be using that instead of the original forward function that NNCF saved during the compressed model creation, you can do this by calling:
185
- model.nncf.set_original_unbound_forward(fn)
186
- if `fn` has an unbound 0-th `self` argument, or
187
- with model.nncf.temporary_bound_original_forward(fn): ...
188
- if `fn` already had 0-th `self` argument bound or never had it in the first place.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
45
  INFO:nncf:Collecting tensor statistics |β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ | 2 / 3
46
  INFO:nncf:Collecting tensor statistics |β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 3 / 3
47
  INFO:nncf:Compiling and loading torch extension: quantized_functions_cuda...
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
48
  INFO:nncf:Finished loading torch extension: quantized_functions_cuda
49
  WARNING:nncf:You are setting `forward` on an NNCF-processed model object.
50
  NNCF relies on custom-wrapping the `forward` call in order to function properly.
 
123
  if `fn` has an unbound 0-th `self` argument, or
124
  with model.nncf.temporary_bound_original_forward(fn): ...
125
  if `fn` already had 0-th `self` argument bound or never had it in the first place.
126
+ INFO:nncf:Statistics of the quantization algorithm:
127
+ Epoch 1 |+--------------------------------+-------+
128
+ Epoch 1 || Statistic's name | Value |
129
+ Epoch 1 |+================================+=======+
130
+ Epoch 1 || Ratio of enabled quantizations | 100 |
131
+ Epoch 1 |+--------------------------------+-------+
132
+ Epoch 1 |
133
+ Epoch 1 |Statistics of the quantization share:
134
+ Epoch 1 |+----------------------------------+--------------------+
135
+ Epoch 1 || Statistic's name | Value |
136
+ Epoch 1 |+==================================+====================+
137
+ Epoch 1 || Symmetric WQs / All placed WQs | 100.00 % (38 / 38) |
138
+ Epoch 1 |+----------------------------------+--------------------+
139
+ Epoch 1 || Asymmetric WQs / All placed WQs | 0.00 % (0 / 38) |
140
+ Epoch 1 |+----------------------------------+--------------------+
141
+ Epoch 1 || Signed WQs / All placed WQs | 100.00 % (38 / 38) |
142
+ Epoch 1 |+----------------------------------+--------------------+
143
+ Epoch 1 || Unsigned WQs / All placed WQs | 0.00 % (0 / 38) |
144
+ Epoch 1 |+----------------------------------+--------------------+
145
+ Epoch 1 || Per-tensor WQs / All placed WQs | 0.00 % (0 / 38) |
146
+ Epoch 1 |+----------------------------------+--------------------+
147
+ Epoch 1 || Per-channel WQs / All placed WQs | 100.00 % (38 / 38) |
148
+ Epoch 1 |+----------------------------------+--------------------+
149
+ Epoch 1 || Placed WQs / Potential WQs | 70.37 % (38 / 54) |
150
+ Epoch 1 |+----------------------------------+--------------------+
151
+ Epoch 1 || Symmetric AQs / All placed AQs | 27.27 % (12 / 44) |
152
+ Epoch 1 |+----------------------------------+--------------------+
153
+ Epoch 1 || Asymmetric AQs / All placed AQs | 72.73 % (32 / 44) |
154
+ Epoch 1 |+----------------------------------+--------------------+
155
+ Epoch 1 || Signed AQs / All placed AQs | 100.00 % (44 / 44) |
156
+ Epoch 1 |+----------------------------------+--------------------+
157
+ Epoch 1 || Unsigned AQs / All placed AQs | 0.00 % (0 / 44) |
158
+ Epoch 1 |+----------------------------------+--------------------+
159
+ Epoch 1 || Per-tensor AQs / All placed AQs | 100.00 % (44 / 44) |
160
+ Epoch 1 |+----------------------------------+--------------------+
161
+ Epoch 1 || Per-channel AQs / All placed AQs | 0.00 % (0 / 44) |
162
+ Epoch 1 |+----------------------------------+--------------------+
163
+ Epoch 1 |
164
+ Epoch 1 |Statistics of the bitwidth distribution:
165
+ Epoch 1 |+--------------+---------------------+--------------------+--------------------+
166
+ Epoch 1 || Num bits (N) | N-bits WQs / Placed | N-bits AQs / | N-bits Qs / Placed |
167
+ Epoch 1 || | WQs | Placed AQs | Qs |
168
+ Epoch 1 |+==============+=====================+====================+====================+
169
+ Epoch 1 || 8 | 100.00 % (38 / 38) | 100.00 % (44 / 44) | 100.00 % (82 / 82) |
170
+ Epoch 1 |+--------------+---------------------+--------------------+--------------------+
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b6286ab040d463864d3a8e9d171df02ea044a878ecd16646b9aef131ecbea578
3
  size 268172814
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:944a5d2394f6acb552e004deceb897c4f1955b3c98a0e7d7582e2f51dbd1e47c
3
  size 268172814
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:77fea7f47d6f91af7370043dfc2a8acdf21fdd576a23651c0002ed739a6042b2
3
  size 4600
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93c59411257f8aad451b10112f68fc1a68f69691452b7210ba820b2e2b131f38
3
  size 4600