• A
  • A
  • A
  • АБВ
  • АБВ
  • АБВ
  • А
  • А
  • А
  • А
  • А
Обычная версия сайта

Семинар HDI Lab: Taming Heterogeneity in Federated Linear Stochastic Approximation

В этот четверг, 05 сентября, в 14:40, состоится доклад Поля Мангольда (Эколь Политехник)

In federated learning, multiple agents collaboratively train a machine learning model without exchanging local data. To achieve this, each agent locally updates a global model, and the updated models are periodically aggregated. In this talk, I will focus on federated linear stochastic approximation (FedLSA), with a strong focus on agents heterogeneity. I will derive upper bounds on the sample and communication complexity of FedLSA, and present a new method to reduce communication cost using control variates. Particular attention will be put on the "linear speed-up" phenomenon, showing that the sample complexity scales with the inverse of the number of agents in both methods.

SCAFFLSA: Taming Heterogeneity in Federated Linear Stochastic Approximation and TD Learning

Доклад состоится дистанционно в Zoom, ссылка для подключения:

Для получения кода доступа к семинару, который пройдет в ZOOM, обращайтесь к Алямовской Елене по указанной почте: ealyamovskaya@hse.ru

Добавить в календарь