admin
commited on
Commit
·
297a25e
1
Parent(s):
dcdd1f4
upd md
Browse files
README.md
CHANGED
@@ -15,7 +15,7 @@ tags:
|
|
15 |
For the 99 recordings, silence is first removed, which is done based on the annotation, targeting the parts where there is no technique annotation. Then all recordings are uniformly segmented into fixed-length segments of 3 seconds. After segmentation, clips shorter than 3 seconds are zero padded. This padding approach, unlike circular padding, is adopted specifically for frame-level detection tasks to prevent the introduction of extraneous information. Regarding the dataset split, since the dataset consists of 99 recordings, we split it at the recording level. The data is partitioned into training, validation, and testing subsets in a 79:10:10 ratio, roughly 8:1:1.
|
16 |
|
17 |
## Demo
|
18 |
-
<https://
|
19 |
|
20 |
## Usage
|
21 |
```python
|
@@ -30,32 +30,17 @@ cd Guzheng_Tech99
|
|
30 |
```
|
31 |
|
32 |
## Results
|
33 |
-
|
|
34 |
-
|
|
35 |
-
|
|
36 |
-
|
|
37 |
| | | | |
|
38 |
-
|
|
39 |
-
| EfficientNet-V2-L |
|
40 |
-
|
|
41 |
-
|
|
42 |
-
|
|
43 |
-
|
|
44 |
-
<!-- Fine-tuning results for a SqueezeNet network on CQT (一个 SqueezeNet 网络在 CQT 上的微调结果):
|
45 |
-
<table>
|
46 |
-
<tr>
|
47 |
-
<th>Loss curve</th>
|
48 |
-
<td><img src="./squeezenet1_1_cqt_2024-07-30_04-38-35/loss.jpg"></td>
|
49 |
-
</tr>
|
50 |
-
<tr>
|
51 |
-
<th>Training and validation accuracy</th>
|
52 |
-
<td><img src="./squeezenet1_1_cqt_2024-07-30_04-38-35/acc.jpg"></td>
|
53 |
-
</tr>
|
54 |
-
<tr>
|
55 |
-
<th>Confusion matrix</th>
|
56 |
-
<td><img src="./squeezenet1_1_cqt_2024-07-30_04-38-35/mat.jpg"></td>
|
57 |
-
</tr>
|
58 |
-
</table> -->
|
59 |
|
60 |
## Dataset
|
61 |
<https://huggingface.co/datasets/ccmusic-database/Guzheng_Tech99>
|
|
|
15 |
For the 99 recordings, silence is first removed, which is done based on the annotation, targeting the parts where there is no technique annotation. Then all recordings are uniformly segmented into fixed-length segments of 3 seconds. After segmentation, clips shorter than 3 seconds are zero padded. This padding approach, unlike circular padding, is adopted specifically for frame-level detection tasks to prevent the introduction of extraneous information. Regarding the dataset split, since the dataset consists of 99 recordings, we split it at the recording level. The data is partitioned into training, validation, and testing subsets in a 79:10:10 ratio, roughly 8:1:1.
|
16 |
|
17 |
## Demo
|
18 |
+
<https://huggingface.co/spaces/ccmusic-database/Guzheng_Tech99>
|
19 |
|
20 |
## Usage
|
21 |
```python
|
|
|
30 |
```
|
31 |
|
32 |
## Results
|
33 |
+
| Backbone | Mel | CQT | Chroma |
|
34 |
+
| :---------------: | :-------: | :-------: | :-------: |
|
35 |
+
| ViT-B-16 | 0.705 | 0.518 | 0.508 |
|
36 |
+
| Swin-T | **0.849** | **0.783** | **0.766** |
|
37 |
| | | | |
|
38 |
+
| VGG19 | **0.862** | 0.799 | 0.665 |
|
39 |
+
| EfficientNet-V2-L | 0.783 | 0.812 | 0.697 |
|
40 |
+
| ConvNeXt-B | 0.849 | **0.849** | **0.805** |
|
41 |
+
| ResNet101 | 0.638 | 0.830 | 0.707 |
|
42 |
+
| SqueezeNet1.1 | 0.831 | 0.814 | 0.780 |
|
43 |
+
| Average | 0.788 | 0.772 | 0.704 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
44 |
|
45 |
## Dataset
|
46 |
<https://huggingface.co/datasets/ccmusic-database/Guzheng_Tech99>
|