Datasets:
The viewer is disabled because this dataset repo requires arbitrary Python code execution. Please consider
removing the
loading script
and relying on
automated data support
(you can use
convert_to_parquet
from the datasets
library). If this is not possible, please
open a discussion
for direct help.
Dataset Card for The Hong Kong Cantonese Corpus (HKCanCor)
Dataset Summary
The Hong Kong Cantonese Corpus (HKCanCor) comprise transcribed conversations recorded between March 1997 and August 1998. It contains recordings of spontaneous speech (51 texts) and radio programmes (42 texts), which involve 2 to 4 speakers, with 1 text of monologue.
In total, the corpus contains around 230,000 Chinese words. The text is word-segmented (i.e., tokenization is at word-level, and each token can span multiple Chinese characters). Tokens are annotated with part-of-speech (POS) tags and romanised Cantonese pronunciation.
- Romanisation
- Follows conventions set by the Linguistic Society of Hong Kong (LSHK).
- POS
- The tagset used by this corpus extends the one in the Peita-Fujitsu-Renmin Ribao (PRF) corpus (Duan et al., 2000). Extensions were made to further capture Cantonese-specific phenomena.
- To facilitate everyday usage and for better comparability across languages and/or corpora, this dataset also includes the tags mapped to the Universal Dependencies 2.0 format. This mapping references the PyCantonese library.
Supported Tasks and Leaderboards
[More Information Needed]
Languages
Yue Chinese / Cantonese (Hong Kong).
Dataset Structure
This corpus has 10801 utterances and approximately 230000 Chinese words. There is no predefined split.
Data Instances
Each instance contains a conversation id, speaker id within that conversation, turn number, part-of-speech tag for each Chinese word in the PRF format and UD2.0 format, and the utterance written in Chinese characters as well as its LSHK format romanisation.
For example:
{
'conversation_id': 'TNR016-DR070398-HAI6V'
'pos_tags_prf': ['v', 'w'],
'pos_tags_ud': ['VERB', 'PUNCT'],
'speaker': 'B',
'transcriptions': ['hai6', 'VQ1'],
'turn_number': 112,
'tokens': ['係', '。']
}
Data Fields
- conversation_id: unique dialogue-level id
- pos_tags_prf: POS tag using the PRF format at token-level
- pos_tag_ud: POS tag using the UD2.0 format at token-level
- speaker: unique speaker id within dialogue
- transcriptions: token-level romanisation in the LSHK format
- turn_number: turn number in dialogue
- tokens: Chinese word or punctuation at token-level
Data Splits
There are no specified splits in this dataset.
Dataset Creation
Curation Rationale
[More Information Needed]
Source Data
Initial Data Collection and Normalization
[More Information Needed]
Who are the source language producers?
[More Information Needed]
Annotations
Annotation process
[More Information Needed]
Who are the annotators?
[More Information Needed]
Personal and Sensitive Information
[More Information Needed]
Considerations for Using the Data
Social Impact of Dataset
[More Information Needed]
Discussion of Biases
[More Information Needed]
Other Known Limitations
[More Information Needed]
Additional Information
Dataset Curators
[More Information Needed]
Licensing Information
This work is licensed under a Creative Commons Attribution 4.0 International License.
Citation Information
This corpus was developed by Luke and Wong, 2015.
@article{luke2015hong,
author={Luke, Kang-Kwong and Wong, May LY},
title={The Hong Kong Cantonese corpus: design and uses},
journal={Journal of Chinese Linguistics},
year={2015},
pages={309-330},
month={12}
}
The POS tagset to Universal Dependency tagset mapping is provided by Jackson Lee, as a part of the PyCantonese library.
@misc{lee2020,
author = {Lee, Jackson},
title = {PyCantonese: Cantonese Linguistics and NLP in Python},
year = {2020},
publisher = {GitHub},
journal = {GitHub repository},
howpublished = {\url{https://github.com/jacksonllee/pycantonese}},
commit = {1d58f44e1cb097faa69de6b617e1d28903b84b98}
}
Contributions
Thanks to @j-chim for adding this dataset.
- Downloads last month
- 80