Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

The Dark Secrets of BERT | Text Machine Blog #1

Open
utterances-bot opened this issue Jan 17, 2020 · 8 comments
Open

The Dark Secrets of BERT | Text Machine Blog #1

utterances-bot opened this issue Jan 17, 2020 · 8 comments

Comments

@utterances-bot
Copy link

The Dark Secrets of BERT | Text Machine Blog

BERT and its Transformer-based cousins are still ahead on all NLP leaderboards. But how much do they actually understand about language?

https://text-machine-lab.github.io/blog/2020/bert-secrets/

Copy link

Fascinating, useful stuff! I have a question about the last study - were attention heads ablated before fine-tuning, so just the LM properties were affected, or after fine-tuning (but before test)?

Copy link

I'm curious to know what the experimental results would be like on the SuperGLUE Benchmark, as it's supposed to be somewhat liberated from the biases and artifacts which are probably being exploited by randomly initialized BERT on the standard GLUE datasets. https://super.gluebenchmark.com/

Copy link

@ruanchaves I agree, that would be indeed an interesting experiment to run, at the time of our submission we only had GLUE available.
@yuvalpinter thanks! the heads were ablated after fine-tuning.

Copy link

Great summarization here.

Copy link

Are your pretrained and finetuned BERT models available for independent analysis ?

Copy link

@charlesmartin14 we haven't released the models themselves, but we consistently used the scripts provided by huggingface to train all of the models.

Copy link

ghk829 commented Apr 29, 2020

Wow the greatest article of the year!
I wonder what’s the black magic!!

Copy link

Thanks a lot for the paper summarization.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

8 participants