KoichiYasuoka's picture
initial release
f59d332
|
raw
history blame
935 Bytes
---
language:
- lzh
tags:
- classical chinese
- literary chinese
- ancient chinese
- token-classification
- pos
- dependency-parsing
base_model: KoichiYasuoka/Xunzi-Qwen2-1.5B-upos
datasets:
- universal_dependencies
license: apache-2.0
pipeline_tag: token-classification
widget:
- text: 子曰學而時習之不亦説乎有朋自遠方來不亦樂乎人不知而不慍不亦君子乎
---
# Xunzi-Qwen2-1.5B-ud-causal
## Model Description
This is a LLaMA model pretrained for POS-tagging and dependency-parsing, derived from [Xunzi-Qwen2-1.5B-upos](https://huggingface.co/KoichiYasuoka/Xunzi-Qwen2-1.5B-upos) and [UD_Classical_Chinese-Kyoto](https://github.com/UniversalDependencies/UD_Classical_Chinese-Kyoto).
## How to Use
```
from transformers import pipeline
nlp=pipeline("universal-dependencies","KoichiYasuoka/Xunzi-Qwen2-1.5B-ud-causal",trust_remote_code=True)
print(nlp("不入虎穴不得虎子"))
```