meeting_2021spring:tlt
Table of Contents
Transfer Learning Theory Reading Group
Location: 1102
Time: Saturday 2pm (every other week)
In this bi-weekly reading group, we will read classic domain adaptation theory papers discussed in the following textbook:
Advances in Domain Adaptation Theory (ADAT), Redko
Through the readings, we hope to get a basic understanding of how and why domain adaptation algorithms work fundamentally, and in what conditions they would not work.
Reading Schedule
Background
- ADAT chapter 1 (Learning theory background)
- ADAT chapter 2 (Domain adaptation background)
Domain adaptation generalization bound
Week 1: HΔH-divergence
- ADAT chapter 3.1-3.4
- Ben-David, Shai, John Blitzer, Koby Crammer, Alex Kulesza, Fernando Pereira, and Jennifer Wortman Vaughan. “A theory of learning from different domains.” Machine learning 79, no. 1 (2010): 151-175 (defines the HΔH-divergence, a preliminary work was published in NIPS 2007)
Week 3: Discrepancy distance I
- ADAT chapter 3.5.1-3.5.2
- Mansour, Yishay, Mehryar Mohri, and Afshin Rostamizadeh. “Domain adaptation: Learning bounds and algorithms.” arXiv preprint arXiv:0902.3430 (2009).improved generalization bound using discrepancy distance
Week 5: Discrepancy distance II
- ADAT chapter 3.5.3 a discrepancy distance based generalization bound for regression problems.
- Cortes, Corinna, and Mehryar Mohri. “Domain adaptation in regression.” In International Conference on Algorithmic Learning Theory, pp. 308-323. Springer, Berlin, Heidelberg, 2011.
- See also:
- Cortes, Corinna, Mehryar Mohri, and Andrés Munoz Medina. “Adaptation based on generalized discrepancy.” The Journal of Machine Learning Research 20, no. 1 (2019): 1-30.
- Maurer, Andreas. “Transfer bounds for linear feature learning.” Machine learning 75, no. 3 (2009): 327-350.
Impossibility theorems for domain adaptation
Week 7: Impossibility theorems
- ADAT Chapter 4.1-4.2
- David, Shai Ben, Tyler Lu, Teresa Luu, and Dávid Pál. “Impossibility theorems for domain adaptation.” In Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics, pp. 129-136. JMLR Workshop and Conference Proceedings, 2010. (using the HΔH-divergence)
Week 9: Hardness results
- ADAT Chapter 4.3-4.4
- Ben-David, Shai, and Ruth Urner. “On the hardness of domain adaptation and the utility of unlabeled target samples.” In International Conference on Algorithmic Learning Theory, pp. 139-153. Springer, Berlin, Heidelberg, 2012.
Integral probability generalization bound
Week 11: Wasserstein distance
- ADAT Chapter 5.1-5.3
- Redko, Ievgen, Amaury Habrard, and Marc Sebban. “Theoretical analysis of domain adaptation with optimal transport.” In Joint European Conference on Machine Learning and Knowledge Discovery in Databases, pp. 737-753. Springer, Cham, 2017.
Other candidate papers
- Baxter, Jonathan. “A model of inductive bias learning.” Journal of artificial intelligence research 12 (2000): 149-198. (Show how multi-task learning can improve generalization, assuming the target task is embedded within an environment of related tasks.)
- ERM-based Multi-source Transfer Learning (recent work by Xinyi on the sample complexity of multi-source transfer learning)
meeting_2021spring/tlt.txt · Last modified: 2021/03/18 22:17 by yang