Cookpadâs experience with self-training BERT with in-house data. BERT with SentencePiece to learn Japanese-specific pre-trained models and solve tasks based on them - Cookpad Developer Blog
-
BERTâs multilingual model is not suitable for handling Japanese, so use SentencePiece.
-
Use Cookpad cooking instructions text (approx. 16 million sentences) for pre-training
-
Learning takes about 3.5 days on a p3.2xlarge instance
- p3.2xlarge is 3USD/hour, so about 25,000.
This page is auto-translated from /nishio/ăŻăăŻăăă+BERT using DeepL. If you looks something interesting but the auto-translated English is not good enough to understand it, feel free to let me know at @nishio_en. Iâm very happy to spread my thought to non-Japanese readers.