Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Docs/SK-646 | Add aggregators and helpers #508

Merged
merged 1 commit into from
Jan 30, 2024
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
38 changes: 38 additions & 0 deletions docs/aggregators.rst
Original file line number Diff line number Diff line change
@@ -0,0 +1,38 @@
Aggregators
===========

Aggregators handle combinations of model updates received by the combiner into a combiner-level global model.
During a training session, the combiners will instantiate an Aggregator and use it to process the incoming model updates from clients.

The above figure illustrates the overall flow. When a client completes a model update, the model parameters are streamed to the combiner, and a model update message is sent. The model parameters are written to file on disk, and the model update message is passed to a callback function, on_model_update. The callback function validates the model update, and if successful, puts the update message on an aggregation queue. The model parameters are written to disk at a configurable storage location at the combiner. This is done to avoid exhausting RAM memory at the combiner. As multiple clients send updates, the aggregation queue builds up, and when a certain criteria is met, another method, combine_models, starts processing the queue, aggregating models according to the specifics of the scheme (FedAvg, FedAdam, etc).

The user can configure several parameters that guide general behavior of the aggregation flow:

- Round timeout: The maximal time the combiner waits before processing the update queue.
- Buffer size: The maximal allowed length of the queue before processing it.
- Whether to retain or delete model update files after they have been processed (default is to delete them)



A developer can extend FEDn with his/her own Aggregator(s) by implementing the interface specified in
:py:mod:`fedn.network.combiners.aggregators.aggregatorbase.AggregatorBase`. The developer implements two following methods:

- ``on_model_update`` (optional)
- ``combine_models``

on_model_update
----------------

The on_model_update has access to the complete model update including the metadata passed on by the clients (as specified in the training entrypoint, see compute package). The base class implements a default callback that checks that all metadata assumed by the aggregation algorithms FedAvg and FedAdam is present in the metadata. However, the callback could also be used to implement custom preprocessing and additional checks including strategies to filter out updates that are suspected to be corrupted or malicious.

combine_models
--------------

This method is responsible for processing the model update queue and in doing so produce an aggregated model. This is the main extension point where the numerical detail of the aggregation scheme is implemented. The best way to understand how to implement this methods is to study the already implemented algorithms:

:py:mod:`fedn.network.combiners.aggregators.fedavg.FedAvg`
:py:mod:`fedn.network.combiners.aggregators.fedopt.FedOpt`

To add an aggregator plugin “myaggregator”, the developer implements the interface and places a file called ‘myaggregator.py’ in the folder ‘fedn.network.combiner.aggregators’.


28 changes: 28 additions & 0 deletions docs/helpers.rst
Original file line number Diff line number Diff line change
@@ -0,0 +1,28 @@
Model Serialization/Deserialization - Helpers
=============================================

In federated learning, model updates need to be serialized and deserialized in order to be
transferred between clients and server/combiner. There is also a need to write and load models
to/from disk, for example to transiently store updates during training rounds.
Furthermore, aggregation algorithms need to perform a range of numerical operations on the
model updates (addition, multiplication, etc). Since different ML frameworks (TF, Torch, etc)
have different internal ways to represent model parameters, there is a need to inform the
framework how to handle models of a given type. In FEDn, this compatibility layer is the
task of Helpers.

A helper is defined by the interface in :py:mod:`fedn.utils.helpers.HelperBase`.
By implementing a helper plugin, a developer can extend the framework with support for new ML
frameworks and numerical operations.

FEDn ships with a default helper implementation, ``numpyhelper``.
This helper relies on the assumption that the model update is made up of parameters
represented by a list of :py:class:`numpy.ndarray` arrays. Since most ML frameworks have
good numpy support it should in most cases be sufficient to use this helper.
Both TF/Keras and PyTorch models can be readily serialized in this way.

To add a helper plugin “myhelper” you implement the interface and place a
file called ‘myhelper.py’ in the folder fedn.utils.helpers.plugins.

See the Keras and PyTorch quickstart examples and :py:mod:`fedn.utils.helpers.plugins.numpyhelper`
for further details.

2 changes: 2 additions & 0 deletions docs/index.rst
Original file line number Diff line number Diff line change
Expand Up @@ -7,6 +7,8 @@
architecture
deployment
interfaces
aggregators
helpers
tutorial
faq
modules
Expand Down
Loading