Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

keras_bert的tokenizer的选取问题 #1

Open
seokjin954 opened this issue May 21, 2021 · 3 comments
Open

keras_bert的tokenizer的选取问题 #1

seokjin954 opened this issue May 21, 2021 · 3 comments

Comments

@seokjin954
Copy link

您好,如果选用项目中的tokenizer,现分词之后会有在分好的词前面有##,且末尾会有[unused1]字符,这样会产生些许差异导致在tokens里找不到sub的token组成的列表,进而无法标记。想请教您是怎么处理的呢

@seokjin954
Copy link
Author

我发现您做的修改了,谢谢

@chenjiawei22
Copy link

你好,请问有ccks2019关系抽取的数据集吗?

@258508
Copy link

258508 commented Sep 7, 2023

求一份数据集,谢谢

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants