This notebook explores using the SpaCy Chinese model to tokenize and analyze parts of speech for a short Buddhist text. Mistakes in the tags were corrected by hand to train an improved model for Pre-Modern Literary Chinese Texts. It was built with Pandas, SpaCy, and Xiaoquan Kong's Chinese SpaCy model. It implements a Part of Speech tagger and v…
-
Updated
Jul 16, 2020 - HTML