Federated Learning (FL) is a new machine learning framework, which enables multiple devices collaboratively to train a shared model without compromising data privacy and security.
This repository will continue to be collected and updated everything about federated learning materials, including research papers, conferences, blogs and beyond.
- Top Machine Learning Conferences
- Top Computer Vision Conferences
- Top Artificial Intelligence and Data Mining Conferences
- Books
- Papers (Research directions)
- Google FL Workshops
- Videos and Lectures
- Tutorials and Blogs
- Open-Sources
In this section, we will summarize Federated Learning papers accepted by top machine learning conference, Including NeurIPS, ICML, ICLR.
In this section, we will summarize Federated Learning papers accepted by top computer vision conference, Including CVPR, ICCV, ECCV.
Years | Title | Affiliation | Materials |
CVPR 2021 | Multi-Institutional Collaborations for Improving Deep Learning-Based Magnetic Resonance Image Reconstruction Using Federated Learning | Johns Hopkins University | code |
Model-Contrastive Federated Learning | National University of Singapore; UC Berkeley |
code | |
FedDG: Federated Domain Generalization on Medical Image Segmentation via Episodic Learning in Continuous Frequency Space | The Chinese University of Hong Kong | code | |
Soteria: Provable Defense Against Privacy Leakage in Federated Learning From Representation Perspective | Duke University | code |
Years | Title | Affiliation | Materials |
ECCV 2020 | Federated Visual Classification with Real-World Data Distribution | MIT; |
Video |
Years | Title | Affiliation | Materials |
ICCV 2021 | Federated Learning for Non-IID Data via Unified Feature Learning and Optimization Objective Alignment | Peking University | |
Ensemble Attention Distillation for Privacy-Preserving Federated Learning | University at Buffalo |
In this section, we will summarize Federated Learning papers accepted by top AI and DM conference, Including AAAI, AISTATS, KDD.
Years | Sessions | Title | Affiliation | Materials |
KDD 2021 | Research Track | Fed2: Feature-Aligned Federated Learning | George Mason University; Microsoft; University of Maryland |
|
FedRS: Federated Learning with Restricted Softmax for Label Distribution Non-IID Data | Nanjing University | |||
Federated Adversarial Debiasing for Fair and Trasnferable Representations | Michigan State University | HomePage | ||
Cross-Node Federated Graph Neural Network for Spatio-Temporal Data Modeling | University of Southern California | code | ||
Application Track | AsySQN: Faster Vertical Federated Learning Algorithms with Better Computation Resource Utilization | |||
FLOP: Federated Learning on Medical Datasets using Partial Networks | Duke University | code | ||
KDD 2020 | Research Track | FedFast: Going Beyond Average for Faster Training of Federated Recommender Systems | University College Dublin | video |
Application Track | Federated Doubly Stochastic Kernel Learning for Vertically Partitioned Data | JD Tech | video |
-
联邦学习(Federated Learning)
-
联邦学习实战(Practicing Federated Learning)
Model Aggregation (or Model Fusion) refers to how to combine local models into a shared global model.
Personalized federated learning refers to train a model for each client, based on the client’s own dataset and the datasets of other clients. There are two major motivations for personalized federated learning:
- Due to statistical heterogeneity across clients, a single global model would not be a good choice for all clients. Sometimes, the local models trained solely on their private data perform better than the global shared model.
- Different clients need models specifically customized to their own environment. As an example of model heterogeneity, consider the sentence: “I live in .....”. The next-word prediction task applied on this sentence needs to predict a different answer customized for each user. Different clients may assign different labels to the same data.
Personalized federated learning Survey paper:
-
Three Approaches for Personalization with Applications to Federated Learning
-
Personalized Federated Learning for Intelligent IoT Applications: A Cloud-Edge based Framework
Recommender system (RecSys) is widely used to solve information overload. In general, the more data RecSys use, the better the recommendation performance we can obtain.
Traditionally, RecSys requires the data that are distributed across multiple devices to be uploaded to the central database for model training. However, due to privacy and security concerns, such directly sharing user data strategies are no longer appropriate.
The incorporation of federated learning and RecSys is a promising approach, which can alleviate the risk of privacy leakage.
Methodology | Papers | Conferences/Affiliations | Materials |
Matrix Factorization | Secure federated matrix factorization | IEEE Intelligent Systems | |
Federated Multi-view Matrix Factorization for Personalized Recommendations | ECML-PKDD 2020 | video | |
Decentralized Recommendation Based on Matrix Factorization: A Comparison of Gossip and Federated Learning | ECML-PKDD 2019 | ||
Towards Privacy-preserving Mobile Applications with Federated Learning: The Case of Matrix Factorization | MobiSys 2019 | ||
Meta Matrix Factorization for Federated Rating Predictions | ACM SIGIR 2020 | code | |
Federated Collaborative Filtering for Privacy-Preserving Personalized Recommendation System | Arxiv | ||
GNN | FedGNN: Federated Graph Neural Network for Privacy-Preserving Recommendation | Arxiv |
Papers | Conferences/Affiliations | Materials |
RPN: A Residual Pooling Network for Efficient Federated Learning | ECAI 2020 (WeBank) |
|
FedBoost: A Communication-Efficient Algorithm for Federated Learning | ICML 2021 (Google) |
Video |
Federated Learning: Strategies for Improving Communication Efficiency | arXiv (Google) |
|
FetchSGD: Communication-Efficient Federated Learning with Sketching | ICML 2020 (UC Berkeley; Johns Hopkins University; Amazon) |
Video Code |
Acceleration for Compressed Gradient Descent in Distributed and Federated Optimization | ICML 2020 (KAUST) |
Slide Video |
Papers | Application Scenarios | Conferences/Affiliations | Materials |
Federated Composite Optimization | loss function contains a non-smooth regularizer | ICML 2021(Google) | code |
Federated Deep AUC Maximization for Hetergeneous Data with a Constant Communication Complexity | Federated Deep AUC Maximization | ICML 2021 (The University of Iowa) |
video |
Bias-Variance Reduced Local SGD for Less Heterogeneous Federated Learning | Non-Convex Objective optimization | ICML 2021 (The University of Tokyo) |
video |
From Local SGD to Local Fixed-Point Methods for Federated Learning | fixed-point algorithms optimization | ICML 2020 (Moscow Institute of Physics and Technology; KAUST) |
Slide Video |
Federated Learning Based on Dynamic Regularization | In each round, the objective function for each device dynamically updates its regularizer | ICLR 2021 (Boston University; ARM) |
|
Federated Learning via Posterior Averaging: A New Perspective and Practical Algorithms | Formulate federated learning optimization as a posterior inference problem | ICLR 2021 (CMU; Google) |
code |
Adaptive Federated Optimization | Federated versions of adaptive optimizers | ICLR 2021 (Google) |
code |
FedBN: Federated Learning on Non-IID Features via Local Batch Normalization | How to uses local batch normalization to alleviate the feature shift before averaging models. | ICLR 2021 (Princeton University) |
code |
Federated Semi-Supervised Learning with Inter-Client Consistency & Disjoint Learning | SFederated versions of emi-Supervised Learning | ICLR 2021 (KAIST) |
code |
Sageflow: Robust Federated Learning against Both Stragglers and Adversaries | Handle both stragglers (slow devices) and adversaries simultaneously | NeurIPS 2021 (KAIST) |
HomePage |
STEM: A Stochastic Two-Sided Momentum Algorithm Achieving Near-Optimal Sample and Communication Complexities for Federated Learning | Distributed stochastic non-convex optimization | University of Minnesota | HomePage |
Secure Bilevel Asynchronous Vertical Federated Learning with Backward Updating | Vertical Federated Learning Optimization | AAAI 2021 (Xidian University; JD Tech) |
video |
Papers | Conferences/Affiliations | Materials |
Addressing Algorithmic Disparity and Performance Inconsistency in Federated Learning | NeurIPS 2021 (Tsinghua University; Alibaba) |
code HomePage |
Fairness-aware Agnostic Federated Learning | SDM 2021 (University of Arkansas) |
|
Fair Resource Allocation in Federated Learning | ICLR 2020 (CMU; Facebook AI) |
Code |
Agnostic Federated Learning | ICML 2019 (Google) |
|
Mitigating Bias in Federated Learning | arXiv (IBM) |
|
Ditto: Fair and Robust Federated Learning Through Personalization | ICML 2021 (CMU; Facebook AI) |
code video |
Category | Papers | Conferences/Affiliations | Materials |
Tree-Base Boosting | Practical Federated Gradient Boosting Decision Trees | AAAI 2020 (NUS) |
code |
Secureboost: A lossless federated learning framework | IEEE Intelligent Systems 2021 (WeBank; HKUST) |
||
Large-scale Secure XGB for Vertical Federated Learning | CIKM 2021 (Ant Group) |
video | |
Papers | Conferences/Affiliations | Materials |
One for One, or All for All: Equilibria and Optimality of Collaboration in Federated Learning | ICML 2021(Toyota; Berkeley; Cornell University) | code video |
Optimality and Stability in Federated Learning: A Game-theoretic Approach | NeurIPS 2021 (Cornell University) |
code HomePage |
Category | Papers | Conferences/Affiliations | Materials |
Clustering | Heterogeneity for the Win: One-Shot Federated Clustering | ICML 2021 (CMU) |
video |
Representations Learning | Exploiting Shared Representations for Personalized Federated Learning | ICML 2021 (University of Texas at Austin; University of Pennsylvania) |
code video |
Papers | Conferences/Affiliations | Materials |
Clustered Sampling: Low-Variance and Improved Representativity for Clients Selection in Federated Learning | ICML 2021 (Accenture) |
code video |
Federated Multi-Armed Bandits | AAAI 2021 (University of Virginia) |
code video |
FedPAQ: A Communication-Efficient Federated Learning Method with Periodic Averaging and Quantization | AISTATS 2020 (UC Santa Barbara; UT Austin) |
video Supplementary |
This section is partially refers to this repository: Awesome-Federated-Learning-on-Graph-and-GNN-papers
-
TensorFlow Federated (TFF): Machine Learning on Decentralized Data - Google, TF Dev Summit ‘19 2019
-
Federated Learning: Machine Learning on Decentralized Data - Google, Google I/O 2019
-
Federated Learning - Cloudera Fast Forward Labs, DataWorks Summit 2019
-
GDPR, Data Shortage and AI - Qiang Yang, AAAI 2019 Invited Talk
- What is Federated Learning - Nvidia 2019
- Online Federated Learning Comic - Google 2019
- Federated Learning: Collaborative Machine Learning without Centralized Training Data - Google AI Blog 2017
- Go Federated with OpenFL - Intel 2021