KennethEnevoldsen commited on
Commit
86656a3
1 Parent(s): 1a7ce16

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +361 -19
README.md CHANGED
@@ -1,4 +1,7 @@
1
  ---
 
 
 
2
  dataset_info:
3
  features:
4
  - name: text
@@ -36,24 +39,363 @@ dataset_info:
36
  - name: dagw_source_full
37
  dtype: string
38
  splits:
39
- - name: train
40
- num_bytes: 4699956.536726911
41
- num_examples: 11762
42
- - name: test
43
- num_bytes: 589804.9133333333
44
- num_examples: 1462
45
  - name: dev
46
- num_bytes: 585061.346
47
- num_examples: 1461
48
- download_size: 1413153
49
- dataset_size: 5874822.796060244
50
- configs:
51
- - config_name: default
52
- data_files:
53
- - split: train
54
- path: data/train-*
55
- - split: test
56
- path: data/test-*
57
- - split: dev
58
- path: data/dev-*
59
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ language: da
3
+ YAML tags:
4
+ - copy-paste the tags obtained with the online tagging app: https://huggingface.co/spaces/huggingface/datasets-tagging
5
  dataset_info:
6
  features:
7
  - name: text
 
39
  - name: dagw_source_full
40
  dtype: string
41
  splits:
 
 
 
 
 
 
42
  - name: dev
43
+ num_bytes: 600679
44
+ num_examples: 1500
45
+ - name: test
46
+ num_bytes: 605135
47
+ num_examples: 1500
48
+ - name: train
49
+ num_bytes: 4819833
50
+ num_examples: 12062
51
+ download_size: 1439625
52
+ dataset_size: 6025647
 
 
 
53
  ---
54
+
55
+
56
+ ## Dataset Description
57
+
58
+ - **Paper:** [If the dataset was introduced by a paper or there was a paper written describing the dataset, add URL here (landing page for Arxiv paper preferred)]()
59
+
60
+ ### Dataset Summary
61
+
62
+ DANSK: Danish Annotations for NLP Specific TasKs is a dataset consisting of texts from multiple domains, sampled from the Danish GigaWord Corpus (DAGW).
63
+ The dataset was created to fill in the gap of Danish NLP datasets from different domains, that are required for training models that generalize across domains. The Named-Entity annotations are moreover fine-grained and have a similar form to that of OntoNotes v5, which significantly broadens the use cases of the dataset.
64
+ The domains include Web, News, Wiki & Books, Legal, Dannet, Conversation and Social Media. For a more in-depth understanding of the domains, please refer to [DAGW](https://huggingface.co/datasets/DDSC/partial-danish-gigaword-no-twitter).
65
+
66
+ The distribution of texts and Named Entities within each domain can be seen in the table below:
67
+
68
+ ### Update log
69
+
70
+ - 2024-03-12: Removed OpenSubtitles from DANSK due to potential copyright infringement
71
+ - 2023-05-26: Added individual annotations for each annotator to allow for analysis of inter-annotator agreement
72
+
73
+
74
+ ### Supported Tasks
75
+
76
+ The DANSK dataset currently only supports Named-Entity Recognition, but additional version releases will contain data for more tasks.
77
+
78
+ ### Languages
79
+
80
+ All texts in the dataset are in Danish.
81
+ Slang from various platforms or dialects may appear, consistent with the domains from which the texts originally have been sampled - e.g. Social Media.
82
+
83
+ ## Dataset Structure
84
+
85
+ ### Data Instances
86
+
87
+ The JSON-formatted data is in the form seen below:
88
+
89
+ ```
90
+ {
91
+ "text": "Aborrer over 2 kg er en uhyre sj\u00e6lden fangst.",
92
+ "ents": [{"start": 13, "end": 17, "label": "QUANTITY"}],
93
+ "sents": [{"start": 0, "end": 45}],
94
+ "tokens": [
95
+ {"id": 0, "start": 0, "end": 7},
96
+ {"id": 1, "start": 8, "end": 12},
97
+ {"id": 2, "start": 13, "end": 14},
98
+ {"id": 3, "start": 15, "end": 17},
99
+ {"id": 4, "start": 18, "end": 20},
100
+ {"id": 5, "start": 21, "end": 23},
101
+ {"id": 6, "start": 24, "end": 29},
102
+ {"id": 7, "start": 30, "end": 37},
103
+ {"id": 8, "start": 38, "end": 44},
104
+ {"id": 9, "start": 44, "end": 45},
105
+ ],
106
+ "spans": {"incorrect_spans": []},
107
+ "dagw_source": "wiki",
108
+ "dagw_domain": "Wiki & Books",
109
+ "dagw_source_full": "Wikipedia",
110
+ }
111
+
112
+ ```
113
+
114
+ ### Data Fields
115
+
116
+ - `text`: The text
117
+ - `ents`: The annotated entities
118
+ - `sents`: The sentences of the text
119
+ - `dagw_source`: Shorthand name of the source from which the text has been sampled in the Danish Gigaword Corpus
120
+ - `dagw_source_full`: Full name of the source from which the text has been sampled in the Danish Gigaword Corpus
121
+ - `dagw_domain`: Name of the domain to which the source adheres to
122
+
123
+ ### Data Splits
124
+
125
+ The data was randomly split up into three distinct partitions; train, dev, as well as a test partition.
126
+ The splits come from the same pool, and there are thus no underlying differences between the sets.
127
+ To see the distribution of named entities, and domains of the different partitions,
128
+ please refer to the paper, or read the superficial statistics provided in the Dataset composition section of this markdown
129
+
130
+ ## Descriptive Statistics
131
+
132
+ ### Dataset Composition
133
+
134
+ Named entity annotation composition across partitions can be seen in the table below:
135
+ | | Full | Train | Validation | Test |
136
+ | :------------: | :---: | :------------: | :----------: | :-----------: |
137
+ | Texts | 15062 | 12062 (80%) | 1500 (10%) | 1500 (10%) |
138
+ | Named entities | 14462 | 11638 (80.47%) | 1327 (9.18%) | 1497 (10.25%) |
139
+ | CARDINAL | 2069 | 1702 (82.26%) | 168 (8.12%) | 226 (10.92%) |
140
+ | DATE | 1756 | 1411 (80.35%) | 182 (10.36%) | 163 (9.28%) |
141
+ | EVENT | 211 | 175 (82.94%) | 19 (9.00%) | 17 (8.06%) |
142
+ | FACILITY | 246 | 200 (81.30%) | 25 (10.16%) | 21 (8.54%) |
143
+ | GPE | 1604 | 1276 (79.55%) | 135 (8.42%) | 193 (12.03%) |
144
+ | LANGUAGE | 126 | 53 (42.06%) | 17 (13.49%) | 56 (44.44%) |
145
+ | LAW | 183 | 148 (80.87%) | 17 (9.29%) | 18 (9.84%) |
146
+ | LOCATION | 424 | 351 (82.78%) | 46 (10.85%) | 27 (6.37%) |
147
+ | MONEY | 714 | 566 (79.27%) | 72 (10.08%) | 76 (10.64%) |
148
+ | NORP | 495 | 405 (81.82%) | 41 (8.28%) | 49 (9.90%) |
149
+ | ORDINAL | 127 | 105 (82.68%) | 11 (8.66%) | 11 (8.66%) |
150
+ | ORGANIZATION | 2507 | 1960 (78.18%) | 249 (9.93%) | 298 (11.87%) |
151
+ | PERCENT | 148 | 123 (83.11%) | 13 (8.78%) | 12 (8.11%) |
152
+ | PERSON | 2133 | 1767 (82.84%) | 191 (8.95%) | 175 (8.20%) |
153
+ | PRODUCT | 763 | 634 (83.09%) | 57 (7.47%) | 72 (9.44%) |
154
+ | QUANTITY | 292 | 242 (82.88%) | 28 (9.59%) | 22 (7.53%) |
155
+ | TIME | 218 | 185 (84.86%) | 18 (8.26%) | 15 (6.88%) |
156
+ | WORK OF ART | 419 | 335 (79.95%) | 38 (9.07%) | 46 (10.98%) |
157
+
158
+ ### Domain distribution
159
+
160
+ Domain and source distribution across partitions can be seen in the table below:
161
+ | Domain | Source | Full | Train | Dev | Test |
162
+ | :----------: | :----------------: | :---: | :---: | :---: | :---: |
163
+ | Conversation | Europa Parlamentet | 206 | 173 | 17 | 16 |
164
+ | Conversation | Folketinget | 23 | 21 | 1 | 1 |
165
+ | Conversation | NAAT | 554 | 431 | 50 | 73 |
166
+ | Conversation | OpenSubtitles* | 377 | 300 | 39 | 38 |
167
+ | Conversation | Spontaneous speech | 489 | 395 | 54 | 40 |
168
+ | Dannet | Dannet | 25 | 18 | 4 | 3 |
169
+ | Legal | Retsinformation.dk | 965 | 747 | 105 | 113 |
170
+ | Legal | Skat.dk | 471 | 364 | 53 | 54 |
171
+ | Legal | Retspraktis | 727 | 579 | 76 | 72 |
172
+ | News | DanAvis | 283 | 236 | 20 | 27 |
173
+ | News | TV2R | 138 | 110 | 16 | 12 |
174
+ | Social Media | hestenettet.dk | 554 | 439 | 51 | 64 |
175
+ | Web | Common Crawl | 8270 | 6661 | 826 | 783 |
176
+ | Wiki & Books | adl | 640 | 517 | 57 | 66 |
177
+ | Wiki & Books | Wikipedia | 279 | 208 | 30 | 41 |
178
+ | Wiki & Books | WikiBooks | 335 | 265 | 36 | 34 |
179
+ | Wiki & Books | WikiSource | 455 | 371 | 43 | 41 |
180
+
181
+ > **Note**: Due to OpenSubtitles potentially containing copyrighted data we have removed it from the dataset.
182
+
183
+ ### Entity Distribution across
184
+
185
+ Domain and named entity distributions for the training set can be seen below:
186
+ | | All domains combined | Conversation | Dannet | Legal | News | Social Media | Web | Wiki and Books |
187
+ | :----------: | :------------------: | :----------: | :----: | :---: | :---: | :----------: | :---: | :------------: |
188
+ | DOCS | 12062 | 1320 | 18 | 1690 | 346 | 439 | 6661 | 1361 |
189
+ | ENTS | 11638 | 1060 | 15 | 1292 | 419 | 270 | 7502 | 883 |
190
+ | CARDINAL | 1702 | 346 | 6 | 95 | 35 | 17 | 1144 | 59 |
191
+ | DATE | 1411 | 113 | 5 | 257 | 40 | 29 | 831 | 126 |
192
+ | EVENT | 175 | 43 | 0 | 1 | 9 | 3 | 106 | 8 |
193
+ | FACILITY | 200 | 2 | 0 | 4 | 18 | 3 | 159 | 10 |
194
+ | GPE | 1276 | 130 | 2 | 60 | 68 | 31 | 846 | 128 |
195
+ | LANGUAGE | 53 | 3 | 0 | 0 | 0 | 0 | 34 | 16 |
196
+ | LAW | 148 | 10 | 0 | 100 | 1 | 0 | 22 | 13 |
197
+ | LOCATION | 351 | 18 | 0 | 1 | 7 | 7 | 288 | 29 |
198
+ | MONEY | 566 | 1 | 0 | 62 | 13 | 18 | 472 | 0 |
199
+ | NORP | 405 | 70 | 0 | 61 | 22 | 1 | 188 | 42 |
200
+ | ORDINAL | 105 | 11 | 0 | 17 | 9 | 2 | 43 | 22 |
201
+ | ORGANIZATION | 1960 | 87 | 0 | 400 | 61 | 39 | 1303 | 58 |
202
+ | PERCENT | 123 | 5 | 0 | 10 | 11 | 0 | 91 | 4 |
203
+ | PERSON | 1767 | 189 | 2 | 194 | 101 | 69 | 970 | 121 |
204
+ | PRODUCT | 634 | 3 | 0 | 10 | 2 | 33 | 581 | 3 |
205
+ | QUANTITY | 242 | 1 | 0 | 9 | 6 | 17 | 188 | 20 |
206
+ | TIME | 185 | 16 | 0 | 5 | 13 | 1 | 144 | 6 |
207
+ | WORK OF ART | 335 | 12 | 0 | 6 | 3 | 0 | 92 | 218 |
208
+
209
+ Domain and named entity distributions for the validation set can be seen below:
210
+ | | Sum | Conversation | Dannet | Legal | News | Social Media | Web | Wiki |
211
+ | :----------: | :---: | :----------: | :----: | :---: | :---: | :----------: | :---: | :---: |
212
+ | DOCS | 1500 | 161 | 4 | 234 | 36 | 51 | 826 | 166 |
213
+ | ENTS | 1497 | 110 | 4 | 171 | 43 | 30 | 983 | 143 |
214
+ | CARDINAL | 226 | 41 | 2 | 19 | 7 | 5 | 139 | 13 |
215
+ | DATE | 163 | 11 | 0 | 27 | 6 | 4 | 89 | 26 |
216
+ | EVENT | 17 | 2 | 0 | 0 | 1 | 0 | 13 | 1 |
217
+ | FACILITY | 21 | 1 | 0 | 0 | 0 | 0 | 16 | 4 |
218
+ | GPE | 193 | 17 | 1 | 8 | 7 | 2 | 131 | 25 |
219
+ | LANGUAGE | 56 | 0 | 0 | 0 | 0 | 0 | 50 | 6 |
220
+ | LAW | 18 | 2 | 0 | 8 | 0 | 0 | 8 | 0 |
221
+ | LOCATION | 27 | 2 | 0 | 1 | 0 | 0 | 21 | 3 |
222
+ | MONEY | 76 | 2 | 0 | 9 | 1 | 6 | 58 | 0 |
223
+ | NORP | 49 | 8 | 0 | 8 | 1 | 2 | 21 | 9 |
224
+ | ORDINAL | 11 | 2 | 0 | 2 | 0 | 1 | 3 | 3 |
225
+ | ORGANIZATION | 298 | 6 | 0 | 68 | 5 | 3 | 212 | 4 |
226
+ | PERCENT | 12 | 0 | 0 | 2 | 0 | 0 | 10 | 0 |
227
+ | PERSON | 175 | 16 | 1 | 16 | 11 | 4 | 96 | 20 |
228
+ | PRODUCT | 72 | 0 | 0 | 0 | 0 | 2 | 69 | 1 |
229
+ | QUANTITY | 22 | 0 | 0 | 1 | 2 | 1 | 17 | 1 |
230
+ | TIME | 15 | 0 | 0 | 0 | 2 | 0 | 13 | 0 |
231
+ | WORK OF ART | 46 | 0 | 0 | 2 | 0 | 0 | 17 | 27 |
232
+
233
+ Domain and named entity distributions for the testing set can be seen below:
234
+ | | Sum | Conversation | Dannet | Legal | News | Social Media | Web | Wiki |
235
+ | :----------: | :---: | :----------: | :----: | :---: | :---: | :----------: | :---: | :---: |
236
+ | DOCS | 1500 | 161 | 4 | 234 | 36 | 51 | 826 | 166 |
237
+ | ENTS | 1497 | 110 | 4 | 171 | 43 | 30 | 983 | 143 |
238
+ | CARDINAL | 226 | 41 | 2 | 19 | 7 | 5 | 139 | 13 |
239
+ | DATE | 163 | 11 | 0 | 27 | 6 | 4 | 89 | 26 |
240
+ | EVENT | 17 | 2 | 0 | 0 | 1 | 0 | 13 | 1 |
241
+ | FACILITY | 21 | 1 | 0 | 0 | 0 | 0 | 16 | 4 |
242
+ | GPE | 193 | 17 | 1 | 8 | 7 | 2 | 131 | 25 |
243
+ | LANGUAGE | 56 | 0 | 0 | 0 | 0 | 0 | 50 | 6 |
244
+ | LAW | 18 | 2 | 0 | 8 | 0 | 0 | 8 | 0 |
245
+ | LOCATION | 27 | 2 | 0 | 1 | 0 | 0 | 21 | 3 |
246
+ | MONEY | 76 | 2 | 0 | 9 | 1 | 6 | 58 | 0 |
247
+ | NORP | 49 | 8 | 0 | 8 | 1 | 2 | 21 | 9 |
248
+ | ORDINAL | 11 | 2 | 0 | 2 | 0 | 1 | 3 | 3 |
249
+ | ORGANIZATION | 298 | 6 | 0 | 68 | 5 | 3 | 212 | 4 |
250
+ | PERCENT | 12 | 0 | 0 | 2 | 0 | 0 | 10 | 0 |
251
+ | PERSON | 175 | 16 | 1 | 16 | 11 | 4 | 96 | 20 |
252
+ | PRODUCT | 72 | 0 | 0 | 0 | 0 | 2 | 69 | 1 |
253
+ | QUANTITY | 22 | 0 | 0 | 1 | 2 | 1 | 17 | 1 |
254
+ | TIME | 15 | 0 | 0 | 0 | 2 | 0 | 13 | 0 |
255
+ | WORK OF ART | 46 | 0 | 0 | 2 | 0 | 0 | 17 | 27 |
256
+
257
+ ## Dataset Creation
258
+
259
+ ### Curation Rationale
260
+
261
+ The dataset is meant to fill in the gap of Danish NLP that up until now
262
+ has been missing a dataset with 1) fine-grained named entity recognition
263
+ labels, and 2) high variance in domain origin of texts. As such, it is the
264
+ intention that DANSK should be employed in training by anyone who wishes
265
+ to create models for NER that are both generalizable across domains and
266
+ fine-grained in their predictions. It may also be utilized to assess across-domain evaluations, in order to unfold any potential domain biases. While
267
+ the dataset currently only entails annotations for named entities, it is the
268
+ intention that future versions of the dataset will feature dependency Parsing,
269
+ pos tagging, and possibly revised NER annotations.
270
+
271
+
272
+
273
+ ### Source Data
274
+
275
+ The data collection, annotation, and normalization steps of the data were extensive.
276
+ As the description is too long for this readme, please refer to the associated paper upon its publication for a full description.
277
+
278
+ #### Initial Data Collection and Normalization
279
+
280
+
281
+ ### Annotations
282
+ #### Annotation process
283
+
284
+ To afford high granularity, the DANSK dataset utilized the annotation standard of OntoNotes 5.0.
285
+ The standard features 18 different named entity types. The full description can be seen in the associated paper.
286
+
287
+ #### Who are the annotators?
288
+ 10 English Linguistics Master’s program students from Aarhus University were employed.
289
+ They worked 10 hours/week for six weeks from October 11, 2021, to November 22, 2021.
290
+ Their annotation tasks included part-of-speech tagging, dependency parsing, and NER annotation.
291
+ Named entity annotations and dependency parsing was done from scratch, while the POS tagging consisted of corrections of silver-standard predictions by an NLP model.
292
+
293
+ ### Annotator Compensation
294
+ 10 English Linguistics Master’s program students from Aarhus University
295
+ were employed. They worked 10 hours/week for six weeks from October 11,
296
+ 2021, to November 22, 2021. Their annotation tasks included
297
+ part-of-speech tagging, dependency parsing, and NER annotation. **Annotators were compensated at the standard rate for students, as determined by the collective agreement of the Danish Ministry of Finance and the Central Organization of Teachers and the
298
+ CO10 Central Organization of 2010 (the CO10 joint agreement), which is 140DKK/hour.** Named
299
+ entity annotations and dependency parsing was done from scratch, while
300
+ the POS tagging consisted of corrections of predictions by an NLP model.
301
+
302
+
303
+ ### Automatic correction
304
+
305
+ During the manual correction of the annotation a series of consistent errors were found. These were corrected using the following Regex patterns (see also the Danish Addendum to the Ontonotes annotation guidelines):
306
+
307
+ <details><summary>Regex Patterns</summary>
308
+ <p>
309
+
310
+ For matching with TIME spans, e.g. [16:30 - 17:30] (TIME):
311
+ ```
312
+ \d{1,2}:\d\d ?[-|\||\/] ?\d
313
+ dag: \d{1,2}
314
+ ```
315
+ For matching with DATE spans, e.g. [1938 - 1992] (DATE):
316
+ ```
317
+ \d{2,4} ?[-|–] ?\d{2,4}
318
+ ```
319
+ For matching companies with A/S og ApS,
320
+ ```
321
+ e.g. [Hansens Skomager A/S] (ORGANIZATION):
322
+ ApS
323
+ A\/S
324
+ ```
325
+
326
+ For matching written numerals, e.g. "en":
327
+ ```
328
+ to | to$|^to| To | To$|^To| TO | TO$|^TO|
329
+ tre | tre$|^tre| Tre | Tre$|^Tre| TRE | TRE$|^TRE|
330
+ fire | fire$|^fire| Fire | Fire$|^Fire| FIRE | FIRE$|^FIRE|
331
+ fem | fem$|^fem| Fem | Fem$|^Fem| FEM | FEM$|^FEM|
332
+ seks | seks$|^seks| Seks | Seks$|^Seks| SEKS | SEKS$|
333
+ ^SYV|
334
+ otte | otte$|^otte| Otte | Otte$|^Otte| OTTE | OTTE$|^OTTE|
335
+ ni | ni$|^ni| Ni | Ni$|^Ni| NI | NI$|^NI|
336
+ ti | ti$|^ti| Ti | Ti$|^Ti| TI | TI$|^TI
337
+ ```
338
+
339
+ For matching "Himlen" or "Himmelen" already annotated
340
+ as LOCATION, e.g. "HIMLEN":
341
+ ```
342
+ [Hh][iI][mM][lL][Ee][Nn]|[Hh][iI][mM][mM][Ee][lL][Ee][Nn]
343
+ ```
344
+
345
+ For matching "Gud" already tagged as PERSON, e.g. "GUD":
346
+ ```
347
+ [Gg][Uu][Dd]
348
+ ```
349
+
350
+ For matching telephone numbers wrongly already
351
+ tagged as CARDINAL, e.g. "20 40 44 30":
352
+ ```
353
+ \d{2} \d{2} \d{2} \d{2}
354
+ \+\d{2} \d{2} ?\d{2} ?\d{2} ?\d{2}$
355
+ \+\d{2} \d{2} ?\d{2} ?\d{2} ?\d{2}$
356
+ \d{4} ?\d{4}$
357
+ ^\d{4} ?\d{4}$
358
+ ```
359
+
360
+ For matching websites already
361
+ wrongly tagged as ORGANIZATION:
362
+ ```
363
+ .dk$|.com$
364
+ ```
365
+
366
+ For matching Hotels and Resorts
367
+ already wrongly tagged as ORGANIZATION:
368
+ ```
369
+ .*[h|H]otel.*|.*[R|r]esort.*
370
+ ```
371
+
372
+ For matching numbers including /
373
+ or :, already wrongly tagged as CARDINAL:
374
+ ```
375
+ \/
376
+ \/
377
+
378
+ -
379
+ ```
380
+
381
+ For matching rights already
382
+ wrongly tagged as LAW:
383
+ ```
384
+ [C|c]opyright
385
+ [®|©]
386
+ [f|F]ortrydelsesret
387
+ [o|O]phavsret$
388
+ enneskeret
389
+ ```
390
+
391
+
392
+ </p>
393
+ </details>
394
+
395
+ ### Licensing Information
396
+
397
+ Creative Commons Attribution-Share Alike 4.0 International license
398
+
399
+ ### Citation Information
400
+ The paper is in progress.
401
+