User Tools

Site Tools


meeting_2021spring:tlt

This is an old revision of the document!


Transfer Learning Theory Reading Group

In this reading group, we will read classic domain adaptation theory papers discussed in the following textbook:

Advances in Domain Adaptation Theory (ADAT), Redko

Reading Schedule

Background

  • ADAT chapter 1 (Learning theory background)
  • ADAT chapter 2 (Domain adaptation background)

Domain adaptation generalization bound

Week 1: HΔH-divergence

  • ADAT chapter 3.1-3.4
  • Ben-David, Shai, John Blitzer, Koby Crammer, Alex Kulesza, Fernando Pereira, and Jennifer Wortman Vaughan. “A theory of learning from different domains.” Machine learning 79, no. 1 (2010): 151-175 (defines the HΔH-divergence, a preliminary work was published in NIPS 2007)

Week 2: Discrepancy distance I

  • ADAT chapter 3.5.1-3.5.2
  • Mansour, Yishay, Mehryar Mohri, and Afshin Rostamizadeh. “Domain adaptation: Learning bounds and algorithms.” arXiv preprint arXiv:0902.3430 (2009).improved generalization bound using discrepancy distance

Week 3: Discrepancy distance II

  • ADAT chapter 3.5.3 a discrepancy distance based generalization bound for regression problems.
  • Cortes, Corinna, and Mehryar Mohri. “Domain adaptation in regression.” In International Conference on Algorithmic Learning Theory, pp. 308-323. Springer, Berlin, Heidelberg, 2011.
  • See also:
    • Cortes, Corinna, Mehryar Mohri, and Andrés Munoz Medina. “Adaptation based on generalized discrepancy.” The Journal of Machine Learning Research 20, no. 1 (2019): 1-30.
    • Maurer, Andreas. “Transfer bounds for linear feature learning.” Machine learning 75, no. 3 (2009): 327-350.

Impossibility theorems for domain adaptation

Week 4: Impossibility theorems

  • ADAT Chapter 4.1-4.2
  • David, Shai Ben, Tyler Lu, Teresa Luu, and Dávid Pál. “Impossibility theorems for domain adaptation.” In Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics, pp. 129-136. JMLR Workshop and Conference Proceedings, 2010. (using the HDH-divergence)

Week 5: Hardness results

  • ADAT Chapter 4.3-4.4
  • Ben-David, Shai, and Ruth Urner. “On the hardness of domain adaptation and the utility of unlabeled target samples.” In International Conference on Algorithmic Learning Theory, pp. 139-153. Springer, Berlin, Heidelberg, 2012.

Integral probability generalization bound

Week 6: Wasserstein distance

  • ADAT Chapter 5.1-5.3
  • Redko, Ievgen, Amaury Habrard, and Marc Sebban. “Theoretical analysis of domain adaptation with optimal transport.” In Joint European Conference on Machine Learning and Knowledge Discovery in Databases, pp. 737-753. Springer, Cham, 2017.

Other candidate papers

  • Baxter, Jonathan. “A model of inductive bias learning.” Journal of artificial intelligence research 12 (2000): 149-198. (Show how multi-task learning, assuming the target task is embedded within an environment of related tasks.)
  • ERM-based Multi-source Transfer Learning (recent work by Xinyi on the sample complexity of multi-source transfer learning)
meeting_2021spring/tlt.1614090424.txt.gz · Last modified: 2021/02/23 09:27 by yang