is a wonderful venue (with good coffee!), and the city is a lot of fun. Everything went smoothly and the
organizers did a great job.
You can get a list of papers that I liked from my Twitter feed, so instead I'd like to discuss some broad themes
I sensed.
- Multitask regularization to mitigate sample complexity in RL. Both in video games and in dialog, it is useful to add extra (auxiliary) tasks in order to accelerate learning.
- Leveraging knowledge and memory. Our current models are powerful function approximators, but in NLP especially we need to go beyond "the current example" in order exhibit competence.
- Gradient descent as inference. Whether it's inpainting with a GAN or BLUE score maximization with an RNN, gradient descent is an unreasonably good inference algorithm.
- Careful initialization is important. I suppose traditional optimization people would say "of course", but we're starting to appreciate the importance of good initialization for deep learning. In particular, start close to linear with eigenvalues close to 1. (Balduzzi et. al. , Poole et. al.)
- Convolutions are as good as, and faster than, recurrent models for NLP. Nice work out of Facebook on causal convolutions for seq2seq. This aligns with my personal experience: we use convolutional NLP models in production for computational performance reasons.
- Neural networks are overparameterized. They can be made much sparser without losing accuracy (Molchanov et. al., Lobacheva et. al.).
- maluuba had the best party. Woot!
Hi Paul,
ReplyDeleteFor your first point - i.e. multitask regularization, could you give some citations both from RL world and dialog world.
Consider slide 84 of http://www.thespermwhale.com/jaseweston/icml2016/icml2016-memnn-tutorial.pdf ... the primary task is story comprehension but adding the additional task of predicting the (exact) teacher response helps.
DeleteI also just noticed slide 27 ... but I'm not familiar with that part.
Great! Thanks
ReplyDelete