Skip to content

Latest commit

 

History

History
48 lines (43 loc) · 3.51 KB

File metadata and controls

48 lines (43 loc) · 3.51 KB
created modified tags type status
2024-10-26T20:03
2024-10-26 22:31
nlp
embedding
transformer
text
sentiment-analysis
semantic
semantic-search
classification
text-classification
question-answering
language
named-entity-recognition
paraphrase-detection
text-summarisation
note
completed

Bidirectional Encoder Representations from Transformers (BERT) is a pretrained encoder-only language model based on the transformer architecture.

The paper originally introducing BERT is BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding (2019).

It is a deep neural network model designed to learn a rich, general representation of language. It is pretrained on BooksCorpus (800M words) and English Wikipedia (2,500M words).

The BERT model is intended to be used as a base model to be extended in order to perform specific downstream NLP tasks such as:

Task Explanation
Sentiment Analysis
Text Classification
Question-answering Given a question and a passage of text, the model is used to identify which part of the passage contains the answer to the question
Named Entity Recognition (NER)
Paraphrase Detection
Semantic Search
Extractive text summarisation
In order to be used in a specific NLP task, BERT can either be fined-tuned (i.e. a single classification layer is added to the end of the BERT model and all of the weights in the entire model are tuned based on the specific downstream task), or a portion of the features in the BERT model are used as is in another model (i.e. BERT provides only a text embedding).

References

Related