I was surprised to find out that some institutions require PhD students to publish a certain number of papers before they can graduate. This is not my view; my goal as a supervisor is to train students to be good scientists, and rigid publication targets are not appropriate for this goal.
I would like neural NLG researchers to focus on more challenging datasets, and make some suggestions.
Seven papers which I blogged or tweeted about in 2020, covering evaluation, safety, engineering and system building, and long-term perspective on NLP. I recommend these to all; they made an impact on me, perhaps they will make an impact on you as well!
I was shocked when a PhD student recently told me that he thought he had to focus on end-to-end neural approaches, because this dominates the conferences he wants to publish in. I’m all for research in end-to-end neural, but fixating on this to the exclusion of everything else is a mistake. Especially since end-to-end neural approaches do not currently work very well.