Federated Learning For Image Classification - Ipynb - Colab
Federated Learning For Image Classification - Ipynb - Colab
ipynb - Colab
Show code
NOTE: This colab has been verified to work with the latest released version of the tensorflow_federated pip package, but the Tensorflow
Federated project is still in pre-release development and may not work on main .
In this tutorial, we use the classic MNIST training example to introduce the Federated Learning (FL) API layer of TFF, tff.learning - a set of
higher-level interfaces that can be used to perform common types of federated learning tasks, such as federated training, against user-supplied
models implemented in TensorFlow.
This tutorial, and the Federated Learning API, are intended primarily for users who want to plug their own TensorFlow models into TFF, treating
the latter mostly as a black box. For a more in-depth understanding of TFF and how to implement your own federated learning algorithms, see
the tutorials on the FC Core API - Custom Federated Algorithms Part 1 and Part 2.
For more on tff.learning , continue with the Federated Learning for Text Generation, tutorial which in addition to covering recurrent models,
also demonstrates loading a pre-trained serialized Keras model for refinement with federated learning combined with evaluation using Keras.
%load_ext tensorboard
import collections
import numpy as np
import tensorflow as tf
import tensorflow_federated as tff
np.random.seed(0)
b'Hello, World!'
In order to facilitate experimentation, we seeded the TFF repository with a few datasets, including a federated version of MNIST that contains a
version of the original NIST dataset that has been re-processed using Leaf so that the data is keyed by the original writer of the digits. Since
each writer has a unique style, this dataset exhibits the kind of non-i.i.d. behavior expected of federated datasets.
The data sets returned by load_data() are instances of tff.simulation.ClientData , an interface that allows you to enumerate the set of
users, to construct a tf.data.Dataset that represents the data of a particular user, and to query the structure of individual elements. Here's
how you can use this interface to explore the content of the data set. Keep in mind that while this interface allows you to iterate over clients ids,
this is only a feature of the simulation data. As you will see shortly, client identities are not used by the federated learning framework - their only
purpose is to allow you to select subsets of the data for simulations.
len(emnist_train.client_ids)
3383
emnist_train.element_type_structure
example_dataset = emnist_train.create_tf_dataset_for_client(
emnist_train.client_ids[0])
example_element = next(iter(example_dataset))
example_element['label'].numpy()
https://colab.research.google.com/github/tensorflow/federated/blob/v0.85.0/docs/tutorials/federated_learning_for_image_classification.ipynb#scrollTo=EsvSXGEMgd9G&printMode=true 1/8
8/16/24, 1:09 PM federated_learning_for_image_classification.ipynb - Colab
First, let's grab a sampling of one client's data to get a feel for the examples on one simulated device. Because the dataset we're using has been
keyed by unique writer, the data of one client represents the handwriting of one person for a sample of the digits 0 through 9, simulating the
unique "usage pattern" of one user.
Now let's visualize the number of examples on each client for each MNIST digit label. In the federated environment, the number of examples on
each client can vary quite a bit, depending on user behavior.
Now let's visualize the mean image per client for each MNIST label. This code will produce the mean of each pixel value for all of the user's
examples for one label. We'll see that one client's mean image for a digit will look different than another client's mean image for the same digit,
due to each person's unique handwriting style. We can muse about how each local training round will nudge the model in a different direction on
each client, as we're learning from that user's own unique data in that local round. Later in the tutorial we'll see how we can take each update to
the model from all the clients and aggregate them together into our new global model, that has learned from each of our client's own unique
data.
# Each client has different mean images, meaning each client will be nudging
# the model in their own directions locally.
for i in range(5):
client_dataset = emnist_train.create_tf_dataset_for_client(
emnist_train.client_ids[i])
plot_data = collections.defaultdict(list)
for example in client_dataset:
plot_data[example['label'].numpy()].append(example['pixels'].numpy())
f = plt.figure(i, figsize=(12, 5))
f.suptitle("Client #{}'s Mean Image Per Label".format(i))
for j in range(10):
mean_img = np.mean(plot_data[j], 0)
plt.subplot(2, 5, j+1)
plt.imshow(mean_img.reshape((28, 28)))
plt.axis('off')
User data can be noisy and unreliably labeled. For example, looking at Client #2's data above, we can see that for label 2, it is possible that there
may have been some mislabeled examples creating a noisier mean image.
NUM_CLIENTS = 10
NUM_EPOCHS = 5
BATCH_SIZE = 20
SHUFFLE_BUFFER = 100
PREFETCH_BUFFER = 10
def preprocess(dataset):
def batch_format_fn(element):
"""Flatten a batch `pixels` and return the features as an `OrderedDict`."""
return collections.OrderedDict(
x=tf.reshape(element['pixels'], [-1, 784]),
y=tf.reshape(element['label'], [-1, 1]))
preprocessed_example_dataset = preprocess(example_dataset)
sample_batch
https://colab.research.google.com/github/tensorflow/federated/blob/v0.85.0/docs/tutorials/federated_learning_for_image_classification.ipynb#scrollTo=EsvSXGEMgd9G&printMode=true 2/8
8/16/24, 1:09 PM federated_learning_for_image_classification.ipynb - Colab
[1],
[4],
[7],
[4],
[2],
[2],
[5],
[4],
[1],
[1],
[0],
[0],
[9]], dtype=int32))])
We have almost all the building blocks in place to construct federated data sets.
One of the ways to feed federated data to TFF in a simulation is simply as a Python list, with each element of the list holding the data of an
individual user, whether as a list or as a tf.data.Dataset . Since we already have an interface that provides the latter, let's use it.
Here's a simple helper function that will construct a list of datasets from the given set of users as an input to a round of training or evaluation.
In a typical federated training scenario, we are dealing with potentially a very large population of user devices, only a fraction of which may be
available for training at a given point in time. This is the case, for example, when the client devices are mobile phones that participate in training
only when plugged into a power source, off a metered network, and otherwise idle.
Of course, we are in a simulation environment, and all the data is locally available. Typically then, when running simulations, we would simply
sample a random subset of the clients to be involved in each round of training, generally different in each round.
That said, as you can find out by studying the paper on the Federated Averaging algorithm, achieving convergence in a system with randomly
sampled subsets of clients in each round can take a while, and it would be impractical to have to run hundreds of rounds in this interactive
tutorial.
What we'll do instead is sample the set of clients once, and reuse the same set across rounds to speed up convergence (intentionally over-
fitting to these few user's data). We leave it as an exercise for the reader to modify this tutorial to simulate random sampling - it is fairly easy to
do (once you do, keep in mind that getting the model to converge may take a while).
sample_clients = emnist_train.client_ids[0:NUM_CLIENTS]
def create_keras_model():
return tf.keras.models.Sequential([
tf.keras.layers.InputLayer(input_shape=(784,)),
tf.keras.layers.Dense(10, kernel_initializer='zeros'),
tf.keras.layers.Softmax(),
])
Note: we do not compile the model yet. The loss, metrics, and optimizers are introduced later.
In order to use any model with TFF, it needs to be wrapped in an instance of the tff.learning.models.VariableModel interface, which exposes
methods to stamp the model's forward pass, metadata properties, etc., similarly to Keras, but also introduces additional elements, such as
ways to control the process of computing federated metrics. Let's not worry about this for now; if you have a Keras model like the one we've just
defined above, you can have TFF wrap it for you by invoking tff.learning.models.from_keras_model , passing the model and a sample data
batch as arguments, as shown below.
def model_fn():
# We _must_ create a new model here, and _not_ capture it from an external
# scope. TFF will call this within different graph contexts.
keras_model = create_keras_model()
return tff.learning.models.from_keras_model(
keras_model,
input_spec=preprocessed_example_dataset.element_spec,
loss=tf.keras.losses.SparseCategoricalCrossentropy(),
metrics=[tf.keras.metrics.SparseCategoricalAccuracy()])
Keep in mind that the argument needs to be a constructor (such as model_fn above), not an already-constructed instance, so that the
construction of your model can happen in a context controlled by TFF (if you're curious about the reasons for this, we encourage you to read the
follow-up tutorial on custom algorithms).
One critical note on the Federated Averaging algorithm below, there are 2 optimizers: a client_optimizer and a server_optimizer. The
client_optimizer is only used to compute local model updates on each client. The server_optimizer applies the averaged update to the global
model at the server. In particular, this means that the choice of optimizer and learning rate used may need to be different than the ones you
have used to train the model on a standard i.i.d. dataset. We recommend starting with regular SGD, possibly with a smaller learning rate than
usual. The learning rate we use has not been carefully tuned, feel free to experiment.
training_process = tff.learning.algorithms.build_weighted_fed_avg(
model_fn,
client_optimizer_fn=lambda: tf.keras.optimizers.SGD(learning_rate=0.02),
server_optimizer_fn=lambda: tf.keras.optimizers.SGD(learning_rate=1.0))
What just happened? TFF has constructed a pair of federated computations and packaged them into a tff.templates.IterativeProcess in
which these computations are available as a pair of properties initialize and next .
https://colab.research.google.com/github/tensorflow/federated/blob/v0.85.0/docs/tutorials/federated_learning_for_image_classification.ipynb#scrollTo=EsvSXGEMgd9G&printMode=true 3/8
8/16/24, 1:09 PM federated_learning_for_image_classification.ipynb - Colab
In a nutshell, federated computations are programs in TFF's internal language that can express various federated algorithms (you can find more
about this in the custom algorithms tutorial). In this case, the two computations generated and packed into iterative_process implement
Federated Averaging.
It is a goal of TFF to define computations in a way that they could be executed in real federated learning settings, but currently only local
execution simulation runtime is implemented. To execute a computation in a simulator, you simply invoke it like a Python function. This default
interpreted environment is not designed for high performance, but it will suffice for this tutorial; we expect to provide higher-performance
simulation runtimes to facilitate larger-scale research in future releases.
Let's start with the initialize computation. As is the case for all federated computations, you can think of it as a function. The computation
takes no arguments, and returns one result - the representation of the state of the Federated Averaging process on the server. While we don't
want to dive into the details of TFF, it may be instructive to see what this state looks like. You can visualize it as follows.
print(training_process.initialize.type_signature.formatted_representation())
( -> <
global_model_weights=<
trainable=<
float32[784,10],
float32[10]
>,
non_trainable=<>
>,
distributor=<>,
client_work=<>,
aggregator=<
value_sum_process=<>,
weight_sum_process=<>
>,
finalizer=<
int64,
float32[784,10],
float32[10]
>
>@SERVER)
While the above type signature may at first seem a bit cryptic, you can recognize that the server state consists of a global_model_weights (the
initial model parameters for MNIST that will be distributed to all devices), some empty parameters (like distributor , which governs the server-
to-client communication) and a finalizer component. This last one governs the logic that the server uses to update its model at the end of a
round, and contains an integer representing how many rounds of FedAvg have occurred.
train_state = training_process.initialize()
The second of the pair of federated computations, next , represents a single round of Federated Averaging, which consists of pushing the
server state (including the model parameters) to the clients, on-device training on their local data, collecting and averaging model updates, and
producing a new updated model at the server.
Conceptually, you can think of next as having a functional type signature that looks as follows.
In particular, one should think about next() not as being a function that runs on a server, but rather being a declarative functional
representation of the entire decentralized computation - some of the inputs are provided by the server ( SERVER_STATE ), but each participating
device contributes its own local dataset.
Let's run a single round of training and visualize the results. We can use the federated data we've already generated above for a sample of
users.
round 1, metrics=OrderedDict([('distributor', ()), ('client_work', OrderedDict([('train', OrderedDict([('sparse_categorical_accuracy', 0.12345679), ('loss', 3.1193733), ('num_exampl
Let's run a few more rounds. As noted earlier, typically at this point you would pick a subset of your simulation data from a new randomly
selected sample of users for each round in order to simulate a realistic deployment in which users continuously come and go, but in this
interactive notebook, for the sake of demonstration we'll just reuse the same users, so that the system converges quickly.
NUM_ROUNDS = 11
for round_num in range(2, NUM_ROUNDS):
result = training_process.next(train_state, federated_train_data)
train_state = result.state
train_metrics = result.metrics
print('round {:2d}, metrics={}'.format(round_num, train_metrics))
round 2, metrics=OrderedDict([('distributor', ()), ('client_work', OrderedDict([('train', OrderedDict([('sparse_categorical_accuracy', 0.14012346), ('loss', 2.9851403), ('num_exampl
round 3, metrics=OrderedDict([('distributor', ()), ('client_work', OrderedDict([('train', OrderedDict([('sparse_categorical_accuracy', 0.1590535), ('loss', 2.8617127), ('num_example
round 4, metrics=OrderedDict([('distributor', ()), ('client_work', OrderedDict([('train', OrderedDict([('sparse_categorical_accuracy', 0.17860082), ('loss', 2.7401376), ('num_exampl
round 5, metrics=OrderedDict([('distributor', ()), ('client_work', OrderedDict([('train', OrderedDict([('sparse_categorical_accuracy', 0.20102881), ('loss', 2.6186547), ('num_exampl
round 6, metrics=OrderedDict([('distributor', ()), ('client_work', OrderedDict([('train', OrderedDict([('sparse_categorical_accuracy', 0.22345679), ('loss', 2.5006158), ('num_exampl
round 7, metrics=OrderedDict([('distributor', ()), ('client_work', OrderedDict([('train', OrderedDict([('sparse_categorical_accuracy', 0.24794239), ('loss', 2.3858356), ('num_exampl
round 8, metrics=OrderedDict([('distributor', ()), ('client_work', OrderedDict([('train', OrderedDict([('sparse_categorical_accuracy', 0.27160493), ('loss', 2.2757034), ('num_exampl
round 9, metrics=OrderedDict([('distributor', ()), ('client_work', OrderedDict([('train', OrderedDict([('sparse_categorical_accuracy', 0.2958848), ('loss', 2.17098), ('num_examples'
round 10, metrics=OrderedDict([('distributor', ()), ('client_work', OrderedDict([('train', OrderedDict([('sparse_categorical_accuracy', 0.3251029), ('loss', 2.072707), ('num_examples
Training loss is decreasing after each round of federated training, indicating the model is converging. There are some important caveats with
these training metrics, however, see the section on Evaluation later in this tutorial.
Let's start by creating the directory and the corresponding summary writer to write the metrics to.
Plot the relevant scalar metrics with the same summary writer.
https://colab.research.google.com/github/tensorflow/federated/blob/v0.85.0/docs/tutorials/federated_learning_for_image_classification.ipynb#scrollTo=EsvSXGEMgd9G&printMode=true 4/8
8/16/24, 1:09 PM federated_learning_for_image_classification.ipynb - Colab
Start TensorBoard with the root log directory specified above. It can take a few seconds for the data to load.
# !rm -R /tmp/logs/scalars/*
In order to view evaluation metrics the same way, you can create a separate eval folder, like "logs/scalars/eval", to write to TensorBoard.
However, tff.learning provides a lower-level model interface, tff.learning.models.VariableModel , that exposes the minimal functionality
necessary for using a model for federated learning. Directly implementing this interface (possibly still using building blocks like
tf.keras.layers ) allows for maximum customization without modifying the internals of the federated learning algorithms.
MnistVariables = collections.namedtuple(
'MnistVariables', 'weights bias num_examples loss_sum accuracy_sum')
Here's a method that creates the variables. For the sake of simplicity, we represent all statistics as tf.float32 , as that will eliminate the need
for type conversions at a later stage. Wrapping variable initializers as lambdas is a requirement imposed by resource variables.
def create_mnist_variables():
return MnistVariables(
weights=tf.Variable(
lambda: tf.zeros(dtype=tf.float32, shape=(784, 10)),
name='weights',
trainable=True),
bias=tf.Variable(
lambda: tf.zeros(dtype=tf.float32, shape=(10)),
name='bias',
trainable=True),
num_examples=tf.Variable(0.0, name='num_examples', trainable=False),
loss_sum=tf.Variable(0.0, name='loss_sum', trainable=False),
accuracy_sum=tf.Variable(0.0, name='accuracy_sum', trainable=False))
With the variables for model parameters and cumulative statistics in place, we can now define the forward pass method that computes loss,
emits predictions, and updates the cumulative statistics for a single batch of input data, as follows.
variables.num_examples.assign_add(num_examples)
variables.loss_sum.assign_add(loss * num_examples)
variables.accuracy_sum.assign_add(accuracy * num_examples)
Next, we define two functions that are related to local metrics, again using TensorFlow.
The first function get_local_unfinalized_metrics returns the unfinalized metric values (in addition to model updates, which are handled
automatically) that are eligible to be aggregated to the server in a federated learning or evaluation process.
def get_local_unfinalized_metrics(variables):
return collections.OrderedDict(
num_examples=[variables.num_examples],
loss=[variables.loss_sum, variables.num_examples],
accuracy=[variables.accuracy_sum, variables.num_examples])
The second function get_metric_finalizers returns an OrderedDict of tf.function s with the same keys (i.e., metric names) as
get_local_unfinalized_metrics . Each tf.function takes in the metric's unfinalized values and computes the finalized metric.
def get_metric_finalizers():
return collections.OrderedDict(
num_examples=tf.function(func=lambda x: x[0]),
loss=tf.function(func=lambda x: x[0] / x[1]),
accuracy=tf.function(func=lambda x: x[0] / x[1]))
https://colab.research.google.com/github/tensorflow/federated/blob/v0.85.0/docs/tutorials/federated_learning_for_image_classification.ipynb#scrollTo=EsvSXGEMgd9G&printMode=true 5/8
8/16/24, 1:09 PM federated_learning_for_image_classification.ipynb - Colab
How the local unfinalized metrics returned by get_local_unfinalized_metrics are aggregated across clients are specified by the
metrics_aggregator parameter when defining the federated learning or evaluation processes. For example, in the
tff.learning.algorithms.build_weighted_fed_avg API (shown in the next section), the default value for metrics_aggregator is
tff.learning.metrics.sum_then_finalize , which first sums the unfinalized metrics from CLIENTS , and then applies the metric finalizers at
SERVER .
With all of the above in place, we are ready to construct a model representation for use with TFF similar to one that's generated for you when
you let TFF ingest a Keras model.
import collections
from collections.abc import Callable
class MnistModel(tff.learning.models.VariableModel):
def __init__(self):
self._variables = create_mnist_variables()
@property
def trainable_variables(self):
return [self._variables.weights, self._variables.bias]
@property
def non_trainable_variables(self):
return []
@property
def local_variables(self):
return [
self._variables.num_examples, self._variables.loss_sum,
self._variables.accuracy_sum
]
@property
def input_spec(self):
return collections.OrderedDict(
x=tf.TensorSpec([None, 784], tf.float32),
y=tf.TensorSpec([None, 1], tf.int32))
@tf.function
def predict_on_batch(self, x, training=True):
del training
return predict_on_batch(self._variables, x)
@tf.function
def forward_pass(self, batch, training=True):
del training
loss, predictions = mnist_forward_pass(self._variables, batch)
num_exmaples = tf.shape(batch['x'])[0]
return tff.learning.models.BatchOutput(
loss=loss, predictions=predictions, num_examples=num_exmaples)
@tf.function
def report_local_unfinalized_metrics(
self) -> collections.OrderedDict[str, list[tf.Tensor]]:
"""Creates an `OrderedDict` of metric names to unfinalized values."""
return get_local_unfinalized_metrics(self._variables)
def metric_finalizers(
self) -> collections.OrderedDict[str, Callable[[list[tf.Tensor]], tf.Tensor]]:
"""Creates an `OrderedDict` of metric names to finalizers."""
return get_metric_finalizers()
@tf.function
def reset_metrics(self):
"""Resets metrics variables to initial value."""
for var in self.local_variables:
var.assign(tf.zeros_like(var))
As you can see, the abstract methods and properties defined by tff.learning.models.VariableModel corresponds to the code snippets in the
preceding section that introduced the variables and defined the loss and statistics.
All state that your model will use must be captured as TensorFlow variables, as TFF does not use Python at runtime (remember your code
should be written such that it can be deployed to mobile devices; see the custom algorithms tutorial for a more in-depth commentary on
the reasons).
Your model should describe what form of data it accepts ( input_spec ), as in general, TFF is a strongly-typed environment and wants to
determine type signatures for all components. Declaring the format of your model's input is an essential part of it.
Although technically not required, we recommend wrapping all TensorFlow logic (forward pass, metric calculations, etc.) as
tf.function s, as this helps ensure the TensorFlow can be serialized, and removes the need for explicit control dependencies.
The above is sufficient for evaluation and algorithms like Federated SGD. However, for Federated Averaging, we need to specify how the model
should train locally on each batch. We will specify a local optimizer when building the Federated Averaging algorithm.
training_process = tff.learning.algorithms.build_weighted_fed_avg(
MnistModel,
client_optimizer_fn=lambda: tf.keras.optimizers.SGD(learning_rate=0.02))
train_state = training_process.initialize()
round 1, metrics=OrderedDict([('distributor', ()), ('client_work', OrderedDict([('train', OrderedDict([('num_examples', 4860.0), ('loss', 3.119374), ('accuracy', 0.12345679)]))])),
https://colab.research.google.com/github/tensorflow/federated/blob/v0.85.0/docs/tutorials/federated_learning_for_image_classification.ipynb#scrollTo=EsvSXGEMgd9G&printMode=true 6/8
8/16/24, 1:09 PM federated_learning_for_image_classification.ipynb - Colab
for round_num in range(2, 11):
result = training_process.next(train_state, federated_train_data)
train_state = result.state
metrics = result.metrics
print('round {:2d}, metrics={}'.format(round_num, metrics))
round 2, metrics=OrderedDict([('distributor', ()), ('client_work', OrderedDict([('train', OrderedDict([('num_examples', 4860.0), ('loss', 2.98514), ('accuracy', 0.14012346)]))])), (
round 3, metrics=OrderedDict([('distributor', ()), ('client_work', OrderedDict([('train', OrderedDict([('num_examples', 4860.0), ('loss', 2.8617127), ('accuracy', 0.1590535)]))])),
round 4, metrics=OrderedDict([('distributor', ()), ('client_work', OrderedDict([('train', OrderedDict([('num_examples', 4860.0), ('loss', 2.740137), ('accuracy', 0.17860082)]))])),
round 5, metrics=OrderedDict([('distributor', ()), ('client_work', OrderedDict([('train', OrderedDict([('num_examples', 4860.0), ('loss', 2.6186547), ('accuracy', 0.20102881)]))])),
round 6, metrics=OrderedDict([('distributor', ()), ('client_work', OrderedDict([('train', OrderedDict([('num_examples', 4860.0), ('loss', 2.5006158), ('accuracy', 0.22345679)]))])),
round 7, metrics=OrderedDict([('distributor', ()), ('client_work', OrderedDict([('train', OrderedDict([('num_examples', 4860.0), ('loss', 2.3858361), ('accuracy', 0.24794239)]))])),
round 8, metrics=OrderedDict([('distributor', ()), ('client_work', OrderedDict([('train', OrderedDict([('num_examples', 4860.0), ('loss', 2.275704), ('accuracy', 0.27160493)]))])),
round 9, metrics=OrderedDict([('distributor', ()), ('client_work', OrderedDict([('train', OrderedDict([('num_examples', 4860.0), ('loss', 2.1709805), ('accuracy', 0.2958848)]))])),
round 10, metrics=OrderedDict([('distributor', ()), ('client_work', OrderedDict([('train', OrderedDict([('num_examples', 4860.0), ('loss', 2.0727067), ('accuracy', 0.3251029)]))])),
To see these metrics within TensorBoard, refer to the steps listed above in "Displaying model metrics in TensorBoard".
keyboard_arrow_down Evaluation
All of our experiments so far presented only federated training metrics - the average metrics over all batches of data trained across all clients in
the round. This introduces the normal concerns about overfitting, especially since we used the same set of clients on each round for simplicity,
but there is an additional notion of overfitting in training metrics specific to the Federated Averaging algorithm. This is easiest to see if we
imagine each client had a single batch of data, and we train on that batch for many iterations (epochs). In this case, the local model will quickly
exactly fit to that one batch, and so the local accuracy metric we average will approach 1.0. Thus, these training metrics can be taken as a sign
that training is progressing, but not much more.
To perform evaluation on federated data, you can construct another federated computation designed for just this purpose, using the
tff.learning.build_federated_evaluation function, and passing in your model constructor as an argument. Note that unlike with Federated
Averaging, where we've used MnistTrainableModel , it suffices to pass the MnistModel . Evaluation doesn't perform gradient descent, and there's
no need to construct optimizers.
For experimentation and research, when a centralized test dataset is available, Federated Learning for Text Generation demonstrates another
evaluation option: taking the trained weights from federated learning, applying them to a standard Keras model, and then simply calling
tf.keras.models.Model.evaluate() on a centralized dataset.
evaluation_process = tff.learning.algorithms.build_fed_eval(MnistModel)
You can inspect the abstract type signature of the evaluation function as follows.
print(evaluation_process.next.type_signature.formatted_representation())
(<
state=<
global_model_weights=<
trainable=<
float32[784,10],
float32[10]
>,
non_trainable=<>
>,
distributor=<>,
client_work=<
<>,
<
num_examples=<
float32
>,
loss=<
float32,
float32
>,
accuracy=<
float32,
float32
>
>
>,
aggregator=<
value_sum_process=<>,
weight_sum_process=<>
>,
finalizer=<>
>@SERVER,
client_data={<
x=float32[?,784],
y=int32[?,1]
>*}@CLIENTS
> -> <
state=<
global_model_weights=<
trainable=<
float32[784,10],
float32[10]
>,
non_trainable=<>
>,
distributor=<>,
client_work=<
<>,
<
num_examples=<
float32
>,
loss=<
float32,
float32
>,
accuracy=<
float32,
Be aware that evaluation process is a tff.lenaring.templates.LearningProcess object. The object has an initialize method that will create
the state, but this will contain an untrained model at first. Using the set_model_weights method, one must insert the weights from the training
state to be evaluated.
evaluation_state = evaluation_process.initialize()
model_weights = training_process.get_model_weights(train_state)
evaluation_state = evaluation_process.set_model_weights(evaluation_state, model_weights)
Now with the evaluation state containing the model weights to be evaluated, we can compute evaluation metrics using evaluation datasets by
calling the next method on the process, just like in training.
https://colab.research.google.com/github/tensorflow/federated/blob/v0.85.0/docs/tutorials/federated_learning_for_image_classification.ipynb#scrollTo=EsvSXGEMgd9G&printMode=true 7/8
8/16/24, 1:09 PM federated_learning_for_image_classification.ipynb - Colab
Here's what we get. Note the numbers look marginally better than what was reported by the last round of training above. By convention, the
training metrics reported by the iterative training process generally reflect the performance of the model at the beginning of the training round,
so the evaluation metrics will always be one step ahead.
str(evaluation_output.metrics)
'OrderedDict([('distributor', ()), ('client_work', OrderedDict([('eval', OrderedDict([('current_round_metrics', OrderedDict([('num_examples', 4860.0), ('loss', 1.6654209), ('accurac
y', 0.3621399)])), ('total rounds metrics', OrderedDict([('num examples', 4860.0), ('loss', 1.6654209), ('accuracy', 0.3621399)]))]))])), ('aggregator', OrderedDict([('mean value',
Now, let's compile a test sample of federated data and rerun evaluation on the test data. The data will come from the same sample of real
users, but from a distinct held-out data set.
len(federated_test_data), federated_test_data[0]
(10,
<_PrefetchDataset element_spec=OrderedDict([('x', TensorSpec(shape=(None, 784), dtype=tf.float32, name=None)), ('y', TensorSpec(shape=(None, 1), dtype=tf.int32, name=None))])>)
str(evaluation_output.metrics)
'OrderedDict([('distributor', ()), ('client_work', OrderedDict([('eval', OrderedDict([('current_round_metrics', OrderedDict([('num_examples', 580.0), ('loss', 1.7750846), ('accurac
y', 0.33620688)])), ('total rounds metrics', OrderedDict([('num examples', 580.0), ('loss', 1.7750846), ('accuracy', 0.33620688)]))]))])), ('aggregator', OrderedDict([('mean value',
This concludes the tutorial. We encourage you to play with the parameters (e.g., batch sizes, number of users, epochs, learning rates, etc.), to
modify the code above to simulate training on random samples of users in each round, and to explore the other tutorials we've developed.
https://colab.research.google.com/github/tensorflow/federated/blob/v0.85.0/docs/tutorials/federated_learning_for_image_classification.ipynb#scrollTo=EsvSXGEMgd9G&printMode=true 8/8