As you might have noticed above, MLEdgeModel instances typically won't be used directly. Instead, they will be used through Edge Predictors, which are lightweight classes that can transform input data into the model's expected input features; and can transform the model's output features into easily usable types. Below are the general steps in implementing Edge predictors:
Defining the Predictor
All Edge predictors must inherit from the IMLPredictor<TOutput> interface. The predictor has a single generic type argument, TOutput, which is a developer-friendly type that is returned when a prediction is made. For example, the MobileNetv2Predictor predictor class which classifies an image uses a tuple for its output type:
// The MobileNetv2 classification predictor returns a tupleclassMobileNetv2Predictor:IMLPredictor<(stringlabel,floatconfidence)>{ ... }
Defining Constructors
Edge predictors should define a static Create method which creates an MLEdgeModel instance by loading the model either from a local file or from NatML Hub. Once created, the predictor can be created using a constructor.
This pattern relies on a constructor that accepts an MLEdgeModel instance:
It is highly recommended to keep the constructor private so that consumers can only create the predictor using the Create method.
Here is a full example of our predictor implementation thus far:
Making Predictions
All Edge predictors must implement a public Predict method which accepts a params MLFeature[] and returns a TOutput. In our case, we have:
Within the Predict method, the predictor should do three things:
Input Checking
The predictor should check that the client has provided the correct number of input features, and that the features have the model's expected types. In our case, we will check that the user passes in an image feature:
If these checks fail, an appropriate exception should be thrown. Do this instead of returning an un-initialized output.
Prediction
To make predictions, the predictor must create MLEdgeFeature instances from input features. Creating an MLEdgeFeature typically requires a corresponding MLFeatureType which dictates any required pre-processing when creating the edge feature. You will typically use the model's input feature types for this purpose:
To check if a feature can be used for Edge predictions, cast it to an IMLEdgeFeature and check that the result of the cast is not null.
Once you have created all the required Edge features, you can then make predictions with the MLEdgeModel:
Marshaling
Once you have output Edge features from the model, you can then marshal the feature data into a more developer-friendly type. This is where most of the heavy-lifting happens in a predictor:
Finally, return your predictor's output:
Disposing the Predictor
All Edge predictors must define a Dispose method, because IMLPredictor implements the IDisposable interface. This method should be used to dispose any explicitly-managed resources used by the predictor. If a predictor does not have any explicitly-managed resources to dispose, then the predictor should hide the Dispose method using interface hiding:
The predictor must notDispose any models provided to it. This is the responsibility of the client.
/// <summary>
/// Create an instance of our predictor
/// </summary>
private MobileNetv2Predictor (MLEdgeModel model) {
...
}
MobileNetv2Predictor.cs
public class MobileNetv2Predictor : IMLPredictor<(string label, string confidence)> {
#region --Client API--
/// <summary>
/// Create a custom predictor.
/// </summary>
public static async Task<MobileNetv2Predictor> Create () {
// Load edge model
var model = await MLEdgeModel.Create(...);
// Create predictor
var predictor = new MobileNetv2Predictor(model);
// Return predictor
return predictor;
}
#endregion
#region --Implementation--
private readonly MLEdgeModel model;
private MobileNetv2Predictor (MLEdgeModel model) {
this.model = model;
}
#endregion
}
/// <summary>
/// Make a prediction with the model.
/// </summary>
/// <param name="inputs">Input feature.</param>
/// <returns>Output label with unnormalized confidence value.</returns>
public (string label, float confidence) Predict (params MLFeature[] inputs);
/// <summary>
/// Make a prediction with the model.
/// </summary>
/// <param name="inputs">Input feature.</param>
/// <returns>Output label with unnormalized confidence value.</returns>
public (string label, float confidence) Predict (params MLFeature[] inputs) {
// Check that the input is an image feature
if (!(inputs[0] is MLImageFeature imageFeature))
throw new InvalidArgumentException(@"Predictor makes predictions on image features");
// ...
}
// Get or create the native feature type which the model expects
MLFeatureType inputType = model.inputs[0];
// Create an Edge feature from the input feature
using MLEdgeFeature edgeFeature = (inputFeature as IMLEdgeFeature).Create(inputType);
// Make a prediction with one or more native input features
using var outputFeatures = model.Predict(edgeFeature);
// Marshal the output feature data into a developer-friendly type
var arrayFeature = new MLArrayFeature<float>(outputFeatures[0]);
// Do stuff with this data...
...
// Create the prediction result from the output data
TOutput result = ...;
// Return it
return result;
// Hide the `Dispose` method so that clients cannot use it directly
void IDisposable.Dispose () { }