Skip to content

Conversation

dcarpintero
Copy link
Contributor

Match tokenizer and model versions:

  • When looking at the papers, the vocab size of 'deberta-v3' is 128k whereas in 'deberta-base' was 50k, and the internal architectures have also some differences. This might lead to issues.
  • Therefore it is proposed to update the tokenizer to 'microsoft/deberta-v3-base'

@MaartenGr
Copy link
Contributor

@jalammar Any preference here for updating the tokenizer?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants