Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Code associated with Figure 3 in the paper #8

Open
guanqun-yang opened this issue May 9, 2021 · 0 comments
Open

Code associated with Figure 3 in the paper #8

guanqun-yang opened this issue May 9, 2021 · 0 comments

Comments

@guanqun-yang
Copy link

From the original paper

We project token-level representations obtained from the BERT embedders onto a 2-dimentional space using t-SNE.

And the paper claims that Figure 3 shows the usefulness of pretraining on OntoNotes by showing more compact clusters. However, as the word embeddings returned by transformer model are contextualized, I am wondering how you get the embeddings of individual tokens in the test set and then apply the t-SNE technique. Do you obtain all of the embeddings and then do the average?

Additionally, I could not find the associated code for visualizing embeddings. Would it be possible the code to obtain Figure 3 provided?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant