- Submission Deadlines:
- 1st place: Siloed Federated Learning for Multi-Centric Histopathology Datasets by Jean Ogier du Terrail (Owkin, Inc.), Mathieu Andreux (Owkin, Inc.), Constance Beguier (Owkin, Inc.), Eric W Tramel (Owkin, Inc.)
- 2nd place: Inverse Distance Aggregation for Federated Learning with Non-IID Data by Yousef Yeganeh (Technical University of Munich) et al. and Fed-BioMed: A general open-source frontend framework for federated learning in healthcare by Santiago Silva (INRIA) et al.
- 3rd place: On the Fairness of Privacy-Preserving Representations in Medical Applications by Mhd Hasan Sarhan (Technical University of Munich) et al.
[October] DCL on MICCAI platform link. Looking forward to a great DCL workshop with you!
[August] The notifications of accepted papers are out! Please note that we need your posters and recordings until September 13th.
[July] Due to several requests, we decided to extend the submission deadline. Please note that the intention of submission still has to be entered into the CMT system until 9th of July 2020 and the full paper until 15th of July 2020. NVIDIA sponsors a Titan RTX for the DCL best paper award!
[June] Submission deadline has been updated. Please submit your paper here. All MICCAI workshops - including DCL - will take place virtually this year.
[May] MICCAI's workshop schedule has been announced. The DCL workshop will take place on October 4 in the afternoon.
[April] Our first keynote speaker is confirmed! Dzung Pham will talk about governance and ethics in the era of distributed learning.
[March] The DCL Workshop has been approved for MICCAI 2020
Deep learning empowers enormous scientific advances, with key applications in healthcare. It has been widely accepted that it is possible to achieve better models with growing amounts of data. However, enabling learning on these huge datasets or training huge models in a timely manner requires to distribute the learning on several devices. One particularity in the medical domain, and in the medical imaging setting is that data sharing across different institutions often becomes impractical due to strict privacy regulations, making the collection of large-scale centralized datasets practically impossible.
Some of the problems, therefore, become: how can we train models in a distributed way on several devices? And is it possible to achieve models as strong as those that can be trained on large centralized datasets without sharing data and breaching the restrictions on privacy and property? Distributed machine learning, including Federated Learning (FL) approaches, could be helpful to solve the latter problem. Different institutions can contribute to building more powerful models by performing collaborative training without sharing any training data. The trained model can be distributed across various institutions but not the actual data. We hope that with FL and other forms of distributed and collaborative learning, the objective of training better, more robust models of higher clinical utility while still protecting the privacy within the data can be achieved.
Through the first MICCAI Workshop on Distributed And Collaborative Learning” (DCL) we aim to provide a discussion forum to compare, evaluate and discuss methodological advancements and ideas around federated, distributed, and collaborative learning schemes that are applicable in the medical domain.
Time Starting (UTC) |
Time Ending (UTC) |
Session title and Summary |
14:00 |
14:05 |
Welcome and opening - Nicola Rieke |
14:05 |
14:45 |
Keynote 1 - Dzung Pham |
14:45 |
15:45 |
Oral Session 1: |
- Weight Erosion: an Update Aggregation Scheme for Personalized Collaborative Machine Learning | ||
15:45 |
16:00 |
Break with poster session (Oral 1 papers) |
16:00 |
16:40 |
Keynote 2 - Jayashree Kalpathy-Cramer |
16:40 |
17:40 |
Oral Session 2: |
- Inverse Distance Aggregation for Federated Learning with Non-IID Data |
||
17:40 |
17:55 |
Break with poster session (Oral 2 papers) |
17:55 |
18:15 |
Best paper award & Concluding remarks |
National Institutes of Health
Topic: Governance & Ethics in the Rise of Distributed Learning
Abstract: Distributed learning techniques have been proposed to satisfy the increasing need for larger and more diverse training data sets while circumventing privacy concerns involved with traditional data sharing. However, such techniques have their own set of potential obstacles that must be surmounted to ensure an ethical and secure treatment of health information. In this talk, I will provide an overview of the current challenges and opportunities in adopting distributed learning in the medical imaging community. Approaches to adversarial attacks, differential privacy, and data encryption will be reviewed. Regulatory issues, standardization, and quality assurance will also be discussed with a focus on describing the major practical considerations in implementing distributed learning collaborations. The early steps of this growing area of research will likely play a significant role in shaping the future of artificial intelligence in medicine.
Harvard Medical School and Massachusetts General Hospital
Topic: Collaborative Deep Learning-opportunities and challenges
Abstract: Deep learning has great potential in medical imaging but concerns about bias, model brittleness and generalizability remain. Training robust deep learning models requires large amounts of diverse, well-annotated patient data. However, imaging data cannot always be shared due to patient privacy concerns and lack of data sharing infrastructure. Collaboratively train deep neural networks on multi-institutional data without sharing patient data is an approach to ensure. Some technical approaches for collaborative learning including federated learning, cyclical weight transfer and split learning. There are a number of technical considerations in developing optimal methods for collaborative learning. We will discuss results from multi-institutional learning for breast density estimation in mammography using open-source frameworks.
Weight Erosion: an Update Aggregation Scheme for Personalized Collaborative Machine Learning, Felix H M Grimberg (EPFL); Mary-Anne Hartley (EPFL); Martin Jaggi (EPFL); Sai Praneeth Karimireddy (EPFL)
Fed-BioMed: A general open-source frontend framework for federated learning in healthcare, Santiago Silva (INRIA); Andre Altmann (Stanford University); Boris Gutman (Illinois Institute of Technology); Marco Lorenzi (INRIA)
On the Fairness of Privacy-Preserving Representations in Medical Applications, Mhd Hasan Sarhan (Technical University of Munich); Nassir Navab (Technical University of Munich); Abouzar Eslami (Carl Zeiss Meditec AG.); Shadi Albarqouni (ETH Zurich)
Siloed Federated Learning for Multi-Centric Histopathology Datasets, Jean Ogier du Terrail (Owkin, Inc.); Mathieu Andreux (Owkin, Inc.); Constance Beguier (Owkin, Inc.); Eric W Tramel (Owkin, Inc.)
Inverse Distance Aggregation for Federated Learning with Non-IID Data, Yousef Yeganeh (Technical University of Munich); Azade Farshad (Technical University of Munich); Nassir Navab (Technical University of Munich); Shadi Albarqouni (ETH Zurich)
Federated Learning for Breast Density Classification: A Real-World Implementation, Holger R Roth (NVIDIA); Ken Chang (Harvard); Praveer Singh (Harvard); Wenqi Li (NVIDIA); Nir Neumark (); Vikash Gupta (OSU); Min Yun (Harvard); Evan Leibovitz (Partners); Bernardo Bizzo (MGH-BWH CCDS); Daguang Xu (NVIDIA Corporation); Jayashree Kalpathy-Cramer (MGH/Harvard Medical School); Alvin Ihsani (NVIDIA ); Yuhong Wen (NVIDIA); Varun Buch (MGH-BWH CCDS); Adam McCarthy (MGH-BWH CCDS); Katharina Hoebel (MIT); Jay Patel (MIT); Ittai Dayan (Partners); Ram Naidu (MGH & BWH Center for Clinical Data Science); Thomas Schultz (Partners); Prerna Dogra (NVIDIA); Andy Feng (NVIDIA); Yan Cheng (NVIDIA); Selnur Erdal (OSU); Richard White (OSU); Keith Dreyer (Partners); Felipe C Kitamura (Diagnósticos da América (DASA)); Meesam Shah (ACR); Laura Coombs (ACR); Ahmed Harouni (NVIDIA); Jeffrey Hawley (OSU); Miao Zhang (Stanford); Liangqiong Qu (Stanford University); Jesse Tetreault (Nvidia); Sharut Gupta (QTIM); Mona Flores (NVIDIA); Daniel Rubin (Stanford University); Jayashree Kalpathy-Cramer (Massachusetts General Hospital)
Federated Gradient Averaging for Multi-Site Training with Momentum-Based Optimizers, Samuel W Remedios (Johns Hopkins University); John Butman (National Institutes of Health); Bennett A Landman (Vanderbilt University); Dzung Pham (National Institutes of Health)
Automated Pancreas Segmentation Using Multi-institutional Collaborative Deep Learning, Chen Shen (Nagoya University); Pochuan Wang (National Taiwan University); Holger R Roth (NVIDIA); Dong Yang (NVIDIA Corporation); Daguang Xu (NVIDIA Corporation); Masahiro Oda (Nagoya University); Kazunari Misawa (Aichi Cancer Center); Po-Ting Chen (National Taiwan University Hospital); Wei-Chih Liao (National Taiwan University Hospital); Kao-Lang Liu (National Taiwan University Hospital); Weichung Wang (National Taiwan University); Kensaku Mori (Nagoya University)
TUM/ETH Zürich
King’s College London
NVIDIA
NVIDIA
NVIDIA
University of Pennsilvania
Vanderbilt University
Verb Surgical
NVIDIA
Amir Alansary, Imperial College London
Reuben Dorent, KCL
Ralf Floca, DKFZ
Mark Graham, KCL
Jonny Hancox, NVIDIA Corporation
Yuankai Huo, Vanderbilt University
Klaus Kades, DKFZ
Jayashree Kalpathy-Cramer, MGH/Harvard Medical School
Sarthak Pati, University of Pennsylvania
G Anthony Reina, Intel Corporation
Daniel Rubin, Stanford University
Wojciech Samek, Fraunhofer HHI
Jonas Scherer, DKFZ
Micah J Sheller, Intel Corporation
Ziyue Xu, NVIDIA Corporation
Dong Yang, NVIDIA Corporation
Maximilian Zenk, DKFZ
NVIDIA sponsors a Titan RTX for the DCL best paper award.
Format: Papers will be submitted electronically following Lecture Notes in Computer Science (LNCS) style of up to 8 + 2 pages (same as MICCAI 2020 format). Submissions exceeding page limit will be rejected without review. Latex style files can be found from Springer, which also contains Word instructions. The file format for submissions is Adobe Portable Document Format (PDF). Other formats will not be accepted.
Double Blind Review: DCL reviewing is double blind. Please review the Anonymity guidelines of MICCAI main conference, and confirm that the author field does not break anonymity.
Paper Submission: DCL uses the CMT system for online submission.
Supplemental Material: Supplemental material submission is optional, following same deadline as the main paper. Contents of the supplemental material would be referred to appropriately in the paper, while reviewers are not obliged to read them.
Submission Originality: Submissions should be original, no paper of substantially similar content should be under peer review or has been accepted for a publication elsewhere (conference/journal, not including archived work).
Proceedings: The proceedings of DCL 2020 will be published as part of the joint MICCAI Workshops proceedings with Springer (LNCS)
Papers will be published as part of the MICCAI Satellite Events joint LNCS proceedings.
All papers will be reviewed following a double-blind review process with at least 2 reviewers per submission.
We follow the MICCAI 2020 guideline regarding arXiv: Reviewers are strongly discouraged to search arXiv for submissions they are responsible to review. Even if they come across this information accidentally, they are discouraged to use the information in formulating their informed review of submissions. arXiv papers are not considered prior work since they have not been peer-reviewed. Therefore, citations to those papers are not required and reviewers are asked to not penalize a paper that fails to cite an arXiv submission.
Each reviewer will be able to cast a score from 1 (lowest) to 5 (highest) and papers with average scores higher than 2.5 will be considered acceptable.
The final decision about acceptance/rejection will be made by the PC member according to ranking, quality and the total number of submissions.
Outstanding papers will be selected for an oral presentation.
We will select reviewers from a pool of reputable researchers in the field who have repeatedly published at venues such as MICCAI, MIDL, CVPR, ICCV, IPMI, and ECCV. The review process will be implemented through the CMT platform. We will use the same system to match papers to the appropriate reviewers.