The 18th Text Analytics Symposium, IEICE Tech. Rep., vol.12, no.178, NLC2021-12, pp. 26-29, Sep. 16, 2021
The 18th Text Analytics Symposium, IEICE Tech. Rep.
Recently, general-purpose language models pre-trained on large corpora such as BERT have been widely used. In Japanese, several pre-trained models based on Wikipedia have been published. On the other hand, general-purpose models may not be sufficiently effective in the financial domain because of the use of specialized phrases. In this study, we construct a pre-training model using a corpus of the financial domain, and evaluate it on a task in the financial domain.
Natural language processing; Language resources; BERT; Financial text;
@inproceedings{Suzuki2021-text18, title={{Performance Validation of Pre-Trained BERT in the Financial Domain [in Japanese]}}, author={Masahiro SUZUMI and Hiroki SAKAJI and Masanori HIRANO and Kiyoshi IZUMI}, booktitle={The 18th Text Analytics Symposium, IEICE Tech. Rep.}, issn={2432-6380}, volume={12}, number={178, NLC2021-12}, pages={26-29}, url={https://www.ieice.org/ken/paper/20210916yCfG/}, year={2021} }