Concrete ML supports a wide range of models through the integration of ONNX nodes. In case a specific ONNX node is missing, developers need to add support for the new ONNX nodes.
The ops_impl.py
file is responsible for implementing the computation of ONNX operators using floating-point arithmetic. The implementation should mirror the behavior of the corresponding ONNX operator precisely. This includes adhering to the expected inputs, outputs, and operational semantics.
Refer to the ONNX documentation to grasp the expected behavior, inputs and outputs of the operator.
After implementing the operator in ops_impl.py
, you need to import it into onnx_utils.py
and map it within the ONNX_OPS_TO_NUMPY_IMPL
dictionary. This mapping is crucial for the framework to recognize and utilize the new operator.
Quantized operators are defined in quantized_ops.py
and are used to handle integer arithmetic. Their implementation is required for the new ONNX to be executed in FHE.
There exist two types of quantized operators:
Univariate Non-Linear Operators: Such operator applies transformation on every element of the input without changing its shape. Sigmoid, Tanh, ReLU are examples of such operation. The sigmoid in this file is simply supported as follows:
Linear Layers: Linear layers like Gemm
and Conv
require specific implementations for integer arithmetic. Please refer to the QuantizedGemm
and QuantizedConv
implementations for reference.
Proper testing is essential to ensure the correctness of the new ONNX node support.
There are many locations where tests can be added:
test_onnx_ops_impl.py
: Tests the implementation of the ONNX node in floating points.
test_quantized_ops.py
: Tests the implementation of the ONNX node in integer arithmetic.
Optional: test_compile_torch.py
: Tests the implementation of a specific torch model that contains the new ONNX operator. The model needs to be added in torch_models.py
.
Finally, update the documentation to reflect the newly supported ONNX node.