You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Enhanced the AdaptiveClassifier to support loading both quantized and unquantized ONNX models, with quantized as the default for improved performance. Updated the README with usage instructions and clarified the behavior of ONNX model selection for saving and loading.
-**Quantized (default)**: INT8 quantized, 4x smaller, ~1.14x faster on ARM, 2-4x faster on x86
245
+
-**Unquantized**: Full precision, maximum accuracy, larger file size
246
+
247
+
By default, models are saved with both versions, and the quantized version is automatically loaded for best performance. Use `prefer_quantized=False` if you need maximum accuracy.
0 commit comments