research reading list
1. Resources
2. Diffusion models
3. NLP
3.1. ASR
3.2. general surveys
3.3. community
- How can we improve peer review in NLP – 2020
- An Adversarial Review of “Adversarial Generation of Natural Language” – Yoav Goldberg blog post from 2017
3.4. generalization
3.5. structured prediction
- Optimal Neural Program Synthesis from Multimodal Specifications – 2020 paper from TAUR lab
- [course] Structured Prediction – Yoav Artzi
- Learning Differentiable Programs with Admissible Neural Heuristics – 2020 cite:shah20_learn_differ_progr_with_admis_neural_heuris
3.6. automata
3.7. multi-modal
- Transformer is All You Need: Multimodal Multitask Learning with a Unified Transformer – 2021
- VideoBERT: A Joint Model for Video and Language Representation Learning – 2019
3.8. grounding
- [course] Language Grounding to Vision and Control – Katerina Fragkiadaki
3.9. machine translation
3.10. attention
- LambdaNetworks: Modeling Long-Range Interactions Without Attention
- Attention is Not Explanation – 2019 cite:jain19_atten_is_not_explan
3.11. language model fine tuning
4. Planning/Navigation
4.1. Transformers
4.2. compositional planners
4.3. NLU
4.4. seq2seq
- Sequence-to-Sequence Model for Trajectory Planning of Nonprehensile Manipulation Including Contact Model
- Vision-and-Language Navigation: Interpreting visually-grounded navigation instructions in real environments – 2018
5. ML
5.1. math
- Parr and Howard 2018 – The Matrix Calculus You Need for Deep Learning cite:parr18_matrix_calcul_you_need_deep_learn
- The Modern Mathematics of Deep Learning
- Deep Learning Goodfellow and Bengio
- Math for Deep Learning – Faisal et al
5.2. intersection with functional programming
- Backprop as Functor:A compositional perspective on supervised learning – 2021
- HaskTorch – Justin Le
5.3. probabilistic programming languages
5.4. opinion
- Machine Learning: The Great Stagnation – Mark Saroufim
5.5. machine learning
5.6. GANS
6. Linguistics
- [course] Computational Semantics – Ellie Pavlick
- 2021 – Comprehension of computer code relies primarily on domain-general executive brain regions cite:ivanova20_compr_comput_code_relies_primar
- A case for deep learning in semantics – 2018
7. Data Science
8. Probability and Stats
- [course] Random – random variables and stats
- Bayesian Epsitemology – In particular, principle of conditionalization
9. HCI
- Joseph Chang – HCI at CMU
10. AI
10.1. TODO On the Opportunities and Risks of Foundational Models – Bommasani et al 2021
10.2. TODO The Scaling Hypothesis
10.3. TODO On the Measure of Intelligence
11. Research practice
12. robotics
- modern robotics – northwestern textbook and course