Skip to content

ONE Release 1.28.0

Compare
Choose a tag to compare
@nnfw-bot nnfw-bot released this 18 Jul 02:39
44fd15b

Release Note 1.28.0

ONE Runtime

Python API

  • Support experimental python API
    • Refer howto document for more details

On-device Training

  • Support on-device training with circle model
    • Training parameter can be passed to onert via onert`s experimental API or loading new model format including training information: circle_plus
    • Trained model can be exported to circle model via experimental API nnfw_train_export_circle
    • Supporting Transfer learning from a pre-trained circle model
  • Introduce circle_plus_gen tool
    • Generates a circle_plus model file with a given training hyperparameters
    • Shows a _circle_plus model details

Runtime configuration API

  • onert supports runtime configuration API for prepare and execution phase via experimental APIs
    • nnfw_set_prepare_config sets configuration for prepare phase, and nnfw_reset_prepare_config resets it to default value
    • nnfw_set_execution_config sets configuration for execution phase, and nnfw_reset_execution_config resets it to default value
    • Supporting prepare phase configuration: prepare execution time profile
    • Supporting execution phase configuration: dump minmax data, dump execution trace, dump execution time
  • Introduce new API to set onert workspace directory: nnfw_set_workspace
    • onert workspace directory is used to store intermediate files during prepare and execution phase

Minmax Recorder

  • Now onert's minmax recorder dumps raw file format instead of HDF5 format
  • onert dumps minmax data into workspace directory

On-device Compilation

  • onert supports full quantization of uint8/int16 type weight and activation.
    • To quantize activation, onert requires minmax data of activation.
  • onert supports on-device code generation for special backend requiring special binary format such as DSP, NPU.
    • Introduce new experimental API for code generation: nnfw_codegen

Type-aware model I/O usage

  • If loaded model is quantized model, onert allows float type I/O buffer
    • onert converts float type input buffer to quantized type internally
    • onert fills float type output buffers by converting quantized type output data to float type internally
  • On multimodel package, onert allows edges between quantized model and float type model