AMBERT: BERT with Multi-Grained Tokenization Achieves SOTA Results on English and Chinese NLU Tasks
Researchers from ByteDance AI Lab have proposed a novel pretrained language model, AMBERT (A Multigrained BERT), which leverages both fine-grained and coarse-grained tokenizations to achieve SOTA performance on English and Chinese language tasks.