Datasets:

Languages:
English
ArXiv:
License:
SeJinPark commited on
Commit
e22656f
·
verified ·
1 Parent(s): 4abbc7b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -11
README.md CHANGED
@@ -24,25 +24,23 @@ size_categories:
24
  - **Point of Contact:** [[email protected]](mailto:[email protected]), [[email protected]](mailto:[email protected])
25
 
26
  ## Dataset Description
27
- This dataset includes manually annotated metadata linking audio files to transcriptions, emotions, and other attributes. The `test_freq.parquet` file contains these links and metadata.
28
 
29
  ### Example Usage
30
  There are 'train', 'test_freq', 'test_rare', 'valid_freq', and 'valid_rare' splits. Below is the example usage.
31
-
32
  ```python
33
  from datasets import load_dataset
34
 
35
- gs = load_dataset("speechcolab/gigaspeech", "valid_freq", use_auth_token=True)
36
 
37
  # see structure
38
- print(gs)
39
 
40
  # load audio sample on the fly
41
- audio_input = gs["valid_freq"][0]["audio"] # first decoded audio sample
42
- transcription = gs["valid_freq"][0]["text"] # first transcription
43
  ```
44
 
45
-
46
  ### Supported Tasks
47
  - `multimodal dialogue generation`
48
  - `automatic-speech-recognition`: The dataset can be used to train a model for Automatic Speech Recognition (ASR).
@@ -52,9 +50,7 @@ transcription = gs["valid_freq"][0]["text"] # first transcription
52
  Multidialog contains audio and transcription data in English.
53
 
54
  ## Dataset Structure
55
-
56
  ### Data Instances
57
-
58
  ```python
59
  {
60
  'conv_id': 't_ffa55df6-114d-4b36-87a1-7af6b8b63d9b',
@@ -71,11 +67,9 @@ Multidialog contains audio and transcription data in English.
71
  'emotion': 'Neutral',
72
  'original_full_path': 'audio/youtube/P0004/YOU0000000315.opus'
73
  }
74
-
75
  ```
76
 
77
  ### Data Fields
78
-
79
  * conv_id (string) - unique identifier for each conversation.
80
  * utterance_id (float) - uterrance index.
81
  * from (string) - who the message is from (human, gpt).
 
24
  - **Point of Contact:** [[email protected]](mailto:[email protected]), [[email protected]](mailto:[email protected])
25
 
26
  ## Dataset Description
27
+ This dataset includes manually annotated metadata linking audio files to transcriptions, emotions, and other attributes.
28
 
29
  ### Example Usage
30
  There are 'train', 'test_freq', 'test_rare', 'valid_freq', and 'valid_rare' splits. Below is the example usage.
 
31
  ```python
32
  from datasets import load_dataset
33
 
34
+ MultiD = load_dataset("IVLLab/MultiDialog", "valid_freq", use_auth_token=True)
35
 
36
  # see structure
37
+ print(MultiD)
38
 
39
  # load audio sample on the fly
40
+ audio_input = MultiD["valid_freq"][0]["audio"] # first decoded audio sample
41
+ transcription = MultiD["valid_freq"][0]["value"] # first transcription
42
  ```
43
 
 
44
  ### Supported Tasks
45
  - `multimodal dialogue generation`
46
  - `automatic-speech-recognition`: The dataset can be used to train a model for Automatic Speech Recognition (ASR).
 
50
  Multidialog contains audio and transcription data in English.
51
 
52
  ## Dataset Structure
 
53
  ### Data Instances
 
54
  ```python
55
  {
56
  'conv_id': 't_ffa55df6-114d-4b36-87a1-7af6b8b63d9b',
 
67
  'emotion': 'Neutral',
68
  'original_full_path': 'audio/youtube/P0004/YOU0000000315.opus'
69
  }
 
70
  ```
71
 
72
  ### Data Fields
 
73
  * conv_id (string) - unique identifier for each conversation.
74
  * utterance_id (float) - uterrance index.
75
  * from (string) - who the message is from (human, gpt).