meeting_2021spring:tlt
This is an old revision of the document!
Table of Contents
Transfer Learning Theory Reading Group
In this reading group, we will read classic domain adaptation theory papers discussed in the following textbook:
Reading Schedule
Background
- ADAT chapter 1 (Learning theory background)
- ADAT chapter 2 (Domain adaptation background)
Domain adaptation generalization bound
Week 1: HΔH-divergence
- ADAT chapter 3.1-3.4
- Ben-David, Shai, John Blitzer, Koby Crammer, Alex Kulesza, Fernando Pereira, and Jennifer Wortman Vaughan. “A theory of learning from different domains.” Machine learning 79, no. 1 (2010): 151-175 (defines the HΔH-divergence, a preliminary work was published in NIPS 2007)
Week 3: Discrepancy distance I
- ADAT chapter 3.5.1-3.5.2
- Mansour, Yishay, Mehryar Mohri, and Afshin Rostamizadeh. “Domain adaptation: Learning bounds and algorithms.” arXiv preprint arXiv:0902.3430 (2009).improved generalization bound using discrepancy distance
Week 5: Discrepancy distance II
- ADAT chapter 3.5.3 a discrepancy distance based generalization bound for regression problems.
- Cortes, Corinna, and Mehryar Mohri. “Domain adaptation in regression.” In International Conference on Algorithmic Learning Theory, pp. 308-323. Springer, Berlin, Heidelberg, 2011.
- See also:
- Cortes, Corinna, Mehryar Mohri, and Andrés Munoz Medina. “Adaptation based on generalized discrepancy.” The Journal of Machine Learning Research 20, no. 1 (2019): 1-30.
- Maurer, Andreas. “Transfer bounds for linear feature learning.” Machine learning 75, no. 3 (2009): 327-350.
Impossibility theorems for domain adaptation
Week 7: Impossibility theorems
- ADAT Chapter 4.1-4.2
- David, Shai Ben, Tyler Lu, Teresa Luu, and Dávid Pál. “Impossibility theorems for domain adaptation.” In Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics, pp. 129-136. JMLR Workshop and Conference Proceedings, 2010. (using the HDH-divergence)
Week 9: Hardness results
- ADAT Chapter 4.3-4.4
- Ben-David, Shai, and Ruth Urner. “On the hardness of domain adaptation and the utility of unlabeled target samples.” In International Conference on Algorithmic Learning Theory, pp. 139-153. Springer, Berlin, Heidelberg, 2012.
Integral probability generalization bound
Week 11: Wasserstein distance
- ADAT Chapter 5.1-5.3
- Redko, Ievgen, Amaury Habrard, and Marc Sebban. “Theoretical analysis of domain adaptation with optimal transport.” In Joint European Conference on Machine Learning and Knowledge Discovery in Databases, pp. 737-753. Springer, Cham, 2017.
Other candidate papers
- Baxter, Jonathan. “A model of inductive bias learning.” Journal of artificial intelligence research 12 (2000): 149-198. (Show how multi-task learning, assuming the target task is embedded within an environment of related tasks.)
- ERM-based Multi-source Transfer Learning (recent work by Xinyi on the sample complexity of multi-source transfer learning)
meeting_2021spring/tlt.1614090486.txt.gz · Last modified: 2021/02/23 09:28 by yang