Neural Networks
Concrete-ML provides simple neural networks models with a Scikit-learn interface through the NeuralNetClassifier
and NeuralNetRegressor
classes.
The neural network models are built with Skorch, which provides a scikit-learn like interface to Torch models (more here).
These models use a stack of linear layers and the activation function and the number of neurons in each layer is configurable. This approach is similar to what is available in Scikit-learn using the MLPClassifier
/MLPRegressor
classes. The built-in, fully connected neural network (FCNN) models train easily with a single call to .fit()
, which will automatically quantize the weights and activations. These models use Quantization Aware Training, allowing good performance for low precision (down to 2-3 bit) weights and activations.
While NeuralNetClassifier
and NeuralNetClassifier
provide scikit-learn like models, their architecture is somewhat restricted in order to make training easy and robust. If you need more advanced models, you can convert custom neural networks as described in the FHE-friendly models documentation.
Good quantization parameter values are critical to make models respect FHE constraints. Weights and activations should be quantized to low precision (e.g. 2-4 bits). Furthermore, in cases of overflow, the sparsity of the network can be tuned as described below.
Example usage
To create an instance of a Fully Connected Neural Network you need to instantiate one of the NeuralNetClassifier
and NeuralNetRegressor
classes and configure a number of parameters that are passed to their constructor. Note that some parameters need to be prefixed by module__
, while others don't. Basically, the parameters that are related to the model, i.e. the underlying nn.Module
, must have the prefix. The parameters that are related to training options do not require the prefix.
The Classifier Comparison notebook shows the behavior of built-in neural networks on several synthetic datasets.
The figure above shows, on the right, the Concrete-ML neural network, trained with Quantization Aware Training, in a FHE-compatible configuration. The figure compares this network to the floating point equivalent, trained with scikit-learn.
Architecture parameters
module__n_layers
: number of layers in the FCNN, must be at least 1. Note that this is the total number of layers. For a single hidden layer NN model, setmodule__n_layers=2
module__n_outputs
: number of outputs (classes or targets)module__input_dim
: dimensionality of the inputmodule__activation_function
: can be one of the Torch activations (e.g. nn.ReLU, see the full list here)
Quantization parameters
n_w_bits
(default 3): number of bits for weightsn_a_bits
(default 3): number of bits for activations and inputsn_accum_bits
(default 8): maximum accumulator bit-width that is desired. The implementation will attempt to keep accumulators under this bit-width through pruning, i.e. setting some weights to zero
Training parameters (from Skorch)
max_epochs
: The number of epochs to train the network (default 10)verbose
: Whether to log loss/metrics during training (default: False)lr
: Learning rate (default 0.001)
Other parameters from skorch are in the Skorch documentation
Advanced parameters
module__n_hidden_neurons_multiplier
: The number of hidden neurons will be automatically set proportional to the dimensionality of the input (i.e. the value formodule__input_dim
). This parameter controls the proportionality factor and is set to 4 by default. This value gives good accuracy while avoiding accumulator overflow. See the pruning and quantization sections for more info.
Network input/output
When you have training data in the form of a NumPy array, and targets in a NumPy 1d array, you can set:
Class weights
You can give weights to each class to use in training. Note that this must be supported by the underlying PyTorch loss function.
Overflow errors
The n_hidden_neurons_multiplier
parameter influences training accuracy as it controls the number of non-zero neurons that are allowed in each layer. Increasing n_hidden_neurons_multiplier
improves accuracy, but should take into account precision limitations to avoid overflow in the accumulator. The default value is a good compromise that avoids overflow, in most cases, but you may want to change the value of this parameter to reduce the breadth of the network if you have overflow errors. A value of 1 should be completely safe with respect to overflow.
Last updated