You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Is your feature request related to a problem? Please describe.
The Huggingface Transformers library is widely used for centralized training due to its integration with many frameworks, features, and ease of use. However, it is not designed for federated learning and has several shortcomings for this task. These include persisting the scheduler and optimizer between rounds, and extracting performance-efficient fine-tuning (PEFT) parameters. While these features are not too difficult to implement, they require boilerplate code that could be avoided.
Describe the solution you'd like
Introduce a new FederatedTrainer class that manages the optimizer, scheduler, and trainer state across federated learning rounds. This class should also handle the correct extraction of PEFT parameters, thereby eliminating the need for users bringing in their Transformers models to implement this code themselves.
The text was updated successfully, but these errors were encountered:
Is your feature request related to a problem? Please describe.
The Huggingface Transformers library is widely used for centralized training due to its integration with many frameworks, features, and ease of use. However, it is not designed for federated learning and has several shortcomings for this task. These include persisting the scheduler and optimizer between rounds, and extracting performance-efficient fine-tuning (PEFT) parameters. While these features are not too difficult to implement, they require boilerplate code that could be avoided.
Describe the solution you'd like
Introduce a new FederatedTrainer class that manages the optimizer, scheduler, and trainer state across federated learning rounds. This class should also handle the correct extraction of PEFT parameters, thereby eliminating the need for users bringing in their Transformers models to implement this code themselves.
The text was updated successfully, but these errors were encountered: