Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add sentence tokenization to process longer texts. #71

Open
wants to merge 1 commit into
base: dev
Choose a base branch
from

Conversation

askonivala
Copy link

The supported sequence length of BERT is up to 512 tokens. Adding a simple sentence tokenization to API would enable users to process longer texts.

@tanmayag78
Copy link

Any other way to handle longer texts as time complexity is higher and it will be inefficient while handling huge text. Like Mitie Ner and Stanford Ner are more efficient for handling longer texts though not as accurate as BERT-NER

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants