iCalendar

7 avr. 2022   11h00 — 13h00

Kilian Fatras MILA, Université McGill, Canada

Pour plus d'informations, visitez: https://cerc-datascience.polymtl.ca/coffee/#

Kilian Fatras

Séminaire hybrique sur Zoom et dans la salle de séminaire du GERAD.

Optimal transport distances have found many applications in machine learning for their capacity to compare non-parametric probability distributions. Yet their algorithmic complexity generally prevents their direct use on large scale datasets. Among the possible strategies to alleviate this issue, practitioners can rely on computing estimates of these distances over minibatches of data. While computationally appealing, we highlight in this talk some limits of this strategy, arguing it can lead to undesirable smoothing effects. As an alternative, we suggest that the same minibatch strategy coupled with unbalanced optimal transport can yield more robust behaviours. We discuss the associated theoretical properties, such as unbiased estimators, existence of gradients and concentration bounds. Our experimental study shows that in challenging problems associated to domain adaptation, the use of unbalanced optimal transport leads to significantly better results, competing with or surpassing recent baselines.

Gabriele Dragotto responsable
Federico Bobbio responsable

Lieu

Salle 4488
Pavillon André-Aisenstadt
Campus de l'Université de Montréal
2920, chemin de la Tour
Montréal QC H3T 1J4
Canada

Organismes associés