-
Notifications
You must be signed in to change notification settings - Fork 2
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Question on Dataset Usage Across Clients in Federated Learning Implementation #4
Comments
Hi, thank you for your attention to our work. Following the regular FL setting, we set class distributions of local clients to be Non-iid, and each local client performs local training on its own dataset. You can check more implementation details in the paper. However, to implement this in our code, we first initialize the whole dataset for each local client as follows: Lines 96 to 99 in 867988c
Then, we redistribute data for each local client based on the categories of local client (e.g., only possess data of old task, possess data of both old and new task, only possess data of new task) as follows: Lines 83 to 91 in 867988c
Hope this answer can help you. |
Thank you very much for your detailed explanation and clarification regarding the dataset usage in federated learning. Your response has greatly enhanced my understanding of the project's implementation and has addressed my queries effectively. I appreciate the time and effort you took to guide me through this aspect of your work. Best regards |
Hello, Thank you for your guidance and insights into this project. I've been closely examining the implementation details and have a question regarding the assignment of classes to clients, specifically the segment of code that involves random selection of classes for each client:
Given this approach, there seems to be a possibility that different clients might be assigned the same class. My concern is about the implication of this scenario for federated learning, where the uniqueness of data across clients is paramount. Could this lead to situations where clients, by chance, end up using identical datasets, contrary to the decentralized and privacy-preserving principles of federated learning? Could you please clarify how the implementation ensures diversity in dataset usage among clients, especially in light of the potential for overlapping class assignments? Thank you for your time and assistance. Best regards, Minkyoon Yoo |
Hello,
First off, I'd like to express my admiration for the work you've done on this federated learning project. It's truly insightful and has been a great resource for me as I delve into federated learning.
I have a question regarding a specific part of the code, particularly about the dataset usage across different clients. From my understanding, one of the fundamental principles of federated learning is that each client trains the model locally using its own distinct dataset. However, as I was reviewing the code, I noticed that it seems every client might be using the same train_dataset:
This observation leads me to wonder if the implementation deviates from federated learning's goal of having each client use a unique dataset. Could you please provide some insights into whether this implementation detail is intentional? Perhaps there's an aspect of the code or the federated learning approach I'm misunderstanding, and I'd be eager to learn more about your design choices in this regard.
Thank you so much for taking the time to address my question amidst your busy schedule. I look forward to your response and learning more about this fascinating project.
Best regards,
Minkyoon Yoo
The text was updated successfully, but these errors were encountered: