KWJA: Kyoto-Waseda Japanese Analyzer1
KWJA is an integrated Japanese text analyzer based on foundation models. KWJA performs many text analysis tasks, including:
- Typo correction
- Sentence segmentation
- Word segmentation
- Word normalization
- Morphological analysis
- Word feature tagging
- Base phrase feature tagging
- NER (Named Entity Recognition)
- Dependency parsing
- Predicate-argument structure (PAS) analysis
- Bridging reference resolution
- Coreference resolution
- Discourse relation analysis
- Python: 3.8+
- Dependencies: See pyproject.toml.
- GPUs with CUDA 11.7 (optional)
Install KWJA with pip:
$ pip install kwja
Perform language analysis with the kwja
command (the result is in the KNP format):
# Analyze a text
$ kwja --text "KWJAは日本語の統合解析ツールです。汎用言語モデルを利用し、様々な言語解析を統一的な方法で解いています。"
# Analyze text files and write the result to a file
$ kwja --filename path/to/file1.txt --filename path/to/file2.txt > path/to/analyzed.knp
# Analyze texts interactively
$ kwja
Please end your input with a new line and type "EOD"
KWJAは日本語の統合解析ツールです。汎用言語モデルを利用し、様々な言語解析を統一的な方法で解いています。
EOD
If you use Windows and PowerShell, you need to set PYTHONUTF8
environment variable to 1
:
> $env:PYTHONUTF8 = "1"
> kwja ...
The output is in the KNP format, which looks like the following:
# S-ID:202210010000-0-0 kwja:1.0.2
* 2D
+ 5D <rel type="=" target="ツール" sid="202210011918-0-0" id="5"/><体言><NE:ARTIFACT:KWJA>
KWJA KWJA KWJA 名詞 6 固有名詞 3 * 0 * 0 <基本句-主辞>
は は は 助詞 9 副助詞 2 * 0 * 0 "代表表記:は/は" <代表表記:は/は>
* 2D
+ 2D <体言>
日本 にほん 日本 名詞 6 地名 4 * 0 * 0 "代表表記:日本/にほん 地名:国" <代表表記:日本/にほん><地名:国><基本句-主辞>
+ 4D <体言><係:ノ格>
語 ご 語 名詞 6 普通名詞 1 * 0 * 0 "代表表記:語/ご 漢字読み:音 カテゴリ:抽象物" <代表表記:語/ご><漢字読み:音><カテゴリ:抽象物><基本句-主辞>
の の の 助詞 9 接続助詞 3 * 0 * 0 "代表表記:の/の" <代表表記:の/の>
...
Here are options for kwja
command:
-
--text
: Text to be analyzed. -
--filename
: Path to a text file to be analyzed. You can specify this option multiple times. -
--model-size
: Model size to be used. Specify one oftiny
,base
(default), andlarge
. -
--device
: Device to be used. Specifycpu
orgpu
. If not specified, the device is automatically selected. -
--typo-batch-size
: Batch size for typo module. -
--senter-batch-size
: Batch size for senter module. -
--seq2seq-batch-size
: Batch size for seq2seq module. -
--char-batch-size
: Batch size for character module. -
--word-batch-size
: Batch size for word module. -
--tasks
: Tasks to be performed. Specify one or more of the following values separated by commas:typo
: Typo correctionsenter
: Sentence segmentationseq2seq
: Word segmentation, Word normalization, Reading prediction, lemmatization, and Canonicalization.char
: Word segmentation and Word normalizationword
: Morphological analysis, Named entity recognition, Word feature tagging, Dependency parsing, PAS analysis, Bridging reference resolution, and Coreference resolution
--config-file
: Path to a custom configuration file.
You can read a KNP format file with rhoknp.
from rhoknp import Document
with open("analyzed.knp") as f:
parsed_document = Document.from_knp(f.read())
For more details about KNP format, see Reference.
Make sure you have kwja
command in your path:
$ which kwja
/path/to/kwja
Install rhoknp:
$ pip install rhoknp
Perform language analysis with the kwja
instance:
from rhoknp import KWJA
kwja = KWJA()
analyzed_document = kwja.apply(
"KWJAは日本語の統合解析ツールです。汎用言語モデルを利用し、様々な言語解析を統一的な方法で解いています。"
)
kwja
can be configured with a configuration file to set the default options.
Check Config file content for details.
On non-Windows systems kwja
follows the
XDG Base Directory Specification
convention for the location of the configuration file.
The configuration dir kwja
uses is itself named kwja
.
In that directory it refers to a file named config.yaml
.
For most people it should be enough to put their config file at ~/.config/kwja/config.yaml
.
You can also provide a configuration file in a non-standard location with an environment variable KWJA_CONFIG_FILE
or a command line option --config-file
.
model_size: base
device: cpu
num_workers: 0
torch_compile: false
typo_batch_size: 1
senter_batch_size: 1
seq2seq_batch_size: 1
char_batch_size: 1
word_batch_size: 1
- typo, senter, character, and word modules
- The performance on each task except typo correction and discourse relation analysis is the mean over all the corpora (KC, KWDLC, Fuman, and WAC) and over three runs with different random seeds.
- We set the learning rate of RoBERTaLARGE (word) to 2e-5 because we failed to fine-tune it with a higher learning rate. Other hyperparameters are the same described in configs, which are tuned for DeBERTaBASE.
- seq2seq module
- The performance on each task is the mean over all the corpora (KC, KWDLC, Fuman, and WAC).
- * denotes results of a single run
- Scores are calculated using a separate script from the character and word modules.
- The performance on each task is the mean over all the corpora (KC, KWDLC, Fuman, and WAC).
Task | Model | ||||||
---|---|---|---|---|---|---|---|
v1.x base ( char, word ) |
v2.x base ( char, word / seq2seq ) |
v1.x large ( char, word ) |
v2.x large ( char, word / seq2seq ) |
||||
Typo Correction | 79.0 | 76.7 | 80.8 | 83.1 | |||
Sentence Segmentation | - | 98.4 | - | 98.6 | |||
Word Segmentation | 98.5 | 98.1 / 98.2* | 98.7 | 98.4 / 98.4* | |||
Word Normalization | 44.0 | 15.3 | 39.8 | 48.6 | |||
Morphological Analysis | POS | 99.3 | 99.4 | 99.3 | 99.4 | ||
sub-POS | 98.1 | 98.5 | 98.2 | 98.5 | |||
conjtype | 99.4 | 99.6 | 99.2 | 99.6 | |||
conjform | 99.5 | 99.7 | 99.4 | 99.7 | |||
reading | 95.5 | 95.4 / 96.2* | 90.8 | 95.6 / 96.8* | |||
lemma | - | - / 97.8* | - | - / 98.1* | |||
canon | - | - / 95.2* | - | - / 95.9* | |||
Named Entity Recognition | 83.0 | 84.6 | 82.1 | 85.9 | |||
Linguistic Feature Tagging | word | 98.3 | 98.6 | 98.5 | 98.6 | ||
base phrase | 86.6 | 93.6 | 86.4 | 93.4 | |||
Dependency Parsing | 92.9 | 93.5 | 93.8 | 93.6 | |||
Pas Analysis | 74.2 | 76.9 | 75.3 | 77.5 | |||
Bridging Reference Resolution | 66.5 | 67.3 | 65.2 | 67.5 | |||
Coreference Resolution | 74.9 | 78.6 | 75.9 | 79.2 | |||
Discourse Relation Analysis | 42.2 | 39.2 | 41.3 | 44.3 |
@InProceedings{Ueda2023a,
author = {Nobuhiro Ueda and Kazumasa Omura and Takashi Kodama and Hirokazu Kiyomaru and Yugo Murawaki and Daisuke Kawahara and Sadao Kurohashi},
title = {KWJA: A Unified Japanese Analyzer Based on Foundation Models},
booktitle = {Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics: System Demonstrations},
year = {2023},
address = {Toronto, Canada},
}
@InProceedings{植田2022,
author = {植田 暢大 and 大村 和正 and 児玉 貴志 and 清丸 寛一 and 村脇 有吾 and 河原 大輔 and 黒橋 禎夫},
title = {KWJA:汎用言語モデルに基づく日本語解析器},
booktitle = {第253回自然言語処理研究会},
year = {2022},
address = {京都},
}
@InProceedings{児玉2023,
author = {児玉 貴志 and 植田 暢大 and 大村 和正 and 清丸 寛一 and 村脇 有吾 and 河原 大輔 and 黒橋 禎夫},
title = {テキスト生成モデルによる日本語形態素解析},
booktitle = {言語処理学会 第29回年次大会},
year = {2023},
address = {沖縄},
}
This software is released under the MIT License, see LICENSE.