Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

So we can not change the word embedding with the pretrained LM? #34

Open
herbertchen1 opened this issue Aug 23, 2018 · 1 comment
Open

Comments

@herbertchen1
Copy link

And training the LM is very hard...?

@rodgzilla
Copy link
Contributor

rodgzilla commented Aug 23, 2018

I do not think that you can change the word embedding easily since its dimension must be the same as the output of each layer, in the case of the pre-trained model 768 (cfg.n_embd).

Training a new language model from scratch is indeed quite expensive and tedious.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants