Skip to content

Look into tokenization #2

@DavidNemeskey

Description

@DavidNemeskey

The original BERT was trained with raw text, and punctuation marks were generally seen attached to words. In emBERT, we take the output of emToken, so punctuation marks are tokens in their own right. This discrepancy might affect performance.

  1. Check if this is really the case. The basic tokenization procedure does split punctuation from the end of words, so the problem might not be as acute as it seems at first sight.
  2. Merge punctuation tokens with the words before sending them to the BERT model.
  3. Alternatively, skip emToken altogether?

Metadata

Metadata

Assignees

Labels

enhancementNew feature or request

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions