Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Sign up
Appearance settings

is there a GPT2 tokenizer optimized for Chinese characters? #739

Unanswered
Jessen-Li asked this question in Q&A
Discussion options

tokenizer = AutoTokenizer.from_pretrained("uer/gpt2-chinese-cluecorpussmall", cache_dir="./gpt2_ch")

debug information shows <class 'transformers.models.bert.tokenization_bert.BertTokenizer'>

model = GPT2LMHeadModel.from_pretrained("uer/gpt2-chinese-cluecorpussmall", cache_dir="./gpt2_ch")

is this a mismatch, how to solve it? since GT2Tokenizer has no "uer/gpt2-chinese-cluecorpussmall"

The response of the trained model contains special tokens like [CLS] and [SEP], just remove them would work?

You must be logged in to vote

Replies: 0 comments

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
1 participant

AltStyle によって変換されたページ (->オリジナル) /