diff --git a/docs/sirider/s1/app-development/README.md b/docs/sirider/s1/app-development/README.md
new file mode 100644
index 000000000..2794a58b0
--- /dev/null
+++ b/docs/sirider/s1/app-development/README.md
@@ -0,0 +1,9 @@
+---
+sidebar_position: 4
+---
+
+# 应用开发
+
+主要介绍上层应用开发,比如 QT, WiringX, Mraa 等
+
+
diff --git a/docs/sirider/s1/app-development/zhouyi_model_zoo.md b/docs/sirider/s1/app-development/zhouyi_model_zoo.md
new file mode 100644
index 000000000..a00a60efe
--- /dev/null
+++ b/docs/sirider/s1/app-development/zhouyi_model_zoo.md
@@ -0,0 +1,266 @@
+---
+sidebar_position: 2
+---
+# 周易 Model Zoo
+
+[周易 Model Zoo](https://github.com/Arm-China/Model_zoo) 仓库提供了一套人工智能模型,供周易 SDK 参考使用。
+
+#### **FTP 模型下载 (推荐 FTP 工具 [FileZilla](https://filezilla-project.org/))**
+ - `Host`: sftp://sftp01.armchina.com
+ - `Account`: zhouyi.armchina
+ - `Password`: 114r3cJd
+
+
+| Model | Framework | Input Shape | Model Source | Quant Model |
+|---------------------------|-----------|---------------------------------------------|---------------------------------------------------------------------------------------------------------------------------------------------------|-------------|
+| mobilenet_v2 | caffe | [1,3,224,224] | https://github.com/shicai/MobileNet-Caffe | No |
+| inception_v4 | tf1 | [1,299,299,3] | https://github.com/tensorflow/models/tree/archive/research/slim/nets | No |
+| deeplab_v2 | onnx | [1, 3, 513, 513] | https://github.com/kazuto1011/deeplab-pytorch | No |
+| mtcnn_o | caffe | [1,3,48,48] | https://github.com/CongWeilin/mtcnn-caffe | No |
+| inception_v4 | caffe | [1,3,299,299] | https://github.com/soeaver/caffe-model/tree/master/cls | No |
+| deepspeech_v2 | onnx | [1,385,161,1] | https://github.com/tensorflow/models/tree/archive/research/deep_speech | No |
+| wavenet | onnx | [1,390,23] | https://github.com/buriburisuri/speech-to-text-wavenet | No |
+| shufflenet_v2 | onnx | [1,3,224,224] | https://github.com/TropComplique/shufflenet-v2-tensorflow | No |
+| mobilenet_v2_ssd | onnx | [1,300,300,3] | https://github.com/tensorflow/models/tree/archive/research/object_detection/models | No |
+| facenet | tf1 | [1, 160, 160, 3] | https://github.com/davidsandberg/facenet | No |
+| yolo_v3_tiny | onnx | [1,416,416,3] | https://github.com/onnx/models/tree/main/vision/object_detection_segmentation/tiny-yolov3 | No |
+| yolo_v3 | caffe | [1, 3, 608, 608] | https://github.com/foss-for-synopsys-dwc-arc-processors/synopsys-caffe-models/tree/master/caffe_models/yolo_v3 | No |
+| yolo_v2_416 | tf1 | [1,416,416,3] | https://github.com/wojciechmo/yolo2 | No |
+| yolo_v2_416 | caffe | [1,3,416,416] | https://github.com/tsingjinyun/caffe-yolov2 | No |
+| yolo_v2_416 | onnx | [1,3,416,416] | https://github.com/wojciechmo/yolo2 | No |
+| inception_v4 | onnx | [1,3,299,299] | https://github.com/tensorflow/models/tree/archive/research/slim/nets | No |
+| mobilenet_v2 | onnx | [1,3,224,224] | https://github.com/tensorflow/models/tree/archive/research/slim/nets/mobilenet | No |
+| faster_rcnn | caffe | [1, 3, 224, 224] | https://github.com/rbgirshick/py-faster-rcnn | No |
+| shufflenet_v2 | tf1 | [1,224,224,3] | https://github.com/TropComplique/shufflenet-v2-tensorflow | No |
+| resnet_v1_101 | caffe | [1,3,224,224] | https://github.com/SnailTyan/caffe-model-zoo | No |
+| mtcnn_p | caffe | [1,3,12,12] | https://github.com/CongWeilin/mtcnn-caffe | No |
+| resnet_v1_50 | tf1 | [1,224,224,3] | https://github.com/tensorflow/models/tree/archive/research/slim/nets | No |
+| alexnet | onnx | [1,1,28,28] | https://github.com/tensorflow/models/tree/archive/research/slim/nets | No |
+| mobilenet_v2_ssd | tf1 | [1,300,300,3] | https://github.com/tensorflow/models/tree/archive/research/object_detection/models | No |
+| inception_v3 | tf1 | [1, 299, 299, 3] | https://github.com/tensorflow/models/tree/archive/research/slim/nets | No |
+| shufflenet_v2 | caffe | [1,3,224,224] | https://github.com/Ewenwan/ShuffleNet-2 | No |
+| deepspeech_v2 | tf1 | [1,385,161,1] | https://github.com/tensorflow/models/tree/archive/research/deep_speech | No |
+| inception_v3 | caffe | [1, 3, 299, 299] | https://github.com/soeaver/caffe-model/tree/master/cls | No |
+| mobilenet_v2_ssd | caffe | [1,3,300,300] | https://github.com/chuanqi305/MobileNet-SSD | No |
+| mtcnn_r | caffe | [1,3,24,24] | https://github.com/CongWeilin/mtcnn-caffe | No |
+| resnet_v1_101 | tf1 | [1,224,224,3] | https://github.com/tensorflow/models/tree/archive/research/slim/nets | No |
+| mobilenet_v2 | tf1 | [1,224,224,3] | https://github.com/tensorflow/models/tree/archive/research/slim/nets/mobilenet | No |
+| efficientnet_b5 | tf1 | [1, 456, 456, 3] | https://github.com/tensorflow/tpu/tree/master/models/official/efficientnet | No |
+| yolo_v3 | tf1 | [1, 416, 416, 3] | https://github.com/qqwweee/keras-yolo3 | No |
+| alexnet | tf1 | [1,28,28,1] | https://github.com/tensorflow/models/tree/archive/research/slim/nets | No |
+| resnet_v1_50 | onnx | [1,3,224,224] | https://github.com/tensorflow/models/tree/archive/research/slim/nets | No |
+| inception_v3 | onnx | [1, 3, 299, 299] | https://github.com/tensorflow/models/tree/archive/research/slim/nets | No |
+| vgg_16 | caffe | [1,3,224,224] | https://gist.github.com/ksimonyan/211839e770f7b538e2d8#file-readme-md | No |
+| resnet_v1_101 | onnx | [1,3,224,224] | https://github.com/tensorflow/models/tree/archive/research/slim/nets | No |
+| vgg_16 | onnx | [1,3,224,224] | https://github.com/tensorflow/models/tree/archive/research/slim/nets | No |
+| wavenet | tf1 | [1,390,23] | https://github.com/buriburisuri/speech-to-text-wavenet | No |
+| alexnet | tflite | [1,28,28,1] | https://github.com/tensorflow/models/tree/archive/research/slim/nets | No |
+| mobilenet_v2 | tflite | [1,224,224,3] | https://github.com/tensorflow/models/tree/archive/research/slim/nets/mobilenet | No |
+| shufflenet_v2 | tflite | [1,224,224,3] | https://github.com/TropComplique/shufflenet-v2-tensorflow | No |
+| inception_v3 | tflite | [1, 299, 299, 3] | https://github.com/tensorflow/models/tree/archive/research/slim/nets | No |
+| inception_v4 | tflite | [1,299,299,3] | https://github.com/tensorflow/models/tree/archive/research/slim/nets | No |
+| resnet_v1_50 | tflite | [1,224,224,3] | https://github.com/tensorflow/models/tree/archive/research/slim/nets | No |
+| resnet_v1_101 | tflite | [1,224,224,3] | https://github.com/tensorflow/models/tree/archive/research/slim/nets | No |
+| vgg_16 | tflite | [1,224,224,3] | https://github.com/tensorflow/models/tree/archive/research/slim/nets | No |
+| alexnet | caffe | [1,3,227,227] | https://github.com/BVLC/caffe/tree/master/models/bvlc_alexnet | No |
+| vgg_16 | tf1 | [1,224,224,3] | https://github.com/tensorflow/models/tree/archive/research/slim/nets | No |
+| resnet_v1_50 | caffe | [1,3,224,224] | https://github.com/SnailTyan/caffe-model-zoo | No |
+| deeplab_v3 | tflite | [1,257,257,3] | https://github.com/tensorflow/models/tree/archive/research/deeplab | No |
+| mobilenet_v3 | tflite | [1,224,224,3] | https://github.com/tensorflow/models/tree/master/research/slim/nets/mobilenet | No |
+| peleenet | caffe | [1,3,224,224] | https://github.com/Robert-JunWang/PeleeNet/tree/master/caffe | No |
+| fcn | caffe | [1,3,224,224] | https://github.com/shelhamer/fcn.berkeleyvision.org/tree/master/voc-fcn8s-atonce | No |
+| deeplab_v3 | onnx | [1,3,513,513] | https://github.com/tensorflow/models/tree/archive/research/deeplab | No |
+| erfnet | caffe | [1,3,512,1024] | https://github.com/Yuelong-Yu/ERFNet-Caffe | No |
+| deeplab_v3 | tf1 | [1,513,513,3] | https://github.com/tensorflow/models/tree/archive/research/deeplab | No |
+| centerface | onnx | [1, 3, 640, 640] | https://github.com/ttanzhiqiang/onnx_tensorrt_project | No |
+| vgg_ssd | caffe | [1, 3, 300, 300] | https://github.com/weiliu89/caffe/tree/ssd | No |
+| icnet | caffe | [1, 3, 1025, 2049] | https://github.com/hszhao/ICNet | No |
+| resnet_34_ssd | onnx | [1, 3, 1200, 1200] | https://github.com/mlcommons/inference/tree/master/vision/classification_and_detection | No |
+| resnet_v2_101 | tf1 | [1, 299, 299, 3] | https://github.com/tensorflow/models/tree/master/research/slim#Pretrained | No |
+| squeezenet | tflite | [1, 224, 224, 3] | https://www.tensorflow.org/lite/guide/hosted_models?hl=zh-cn | No |
+| sphereface | caffe | [1, 3, 112, 96] | https://github.com/openvinotoolkit/open_model_zoo/blob/master/models/public/Sphereface/README.md | No |
+| dpn_92 | caffe | [1, 3, 224, 224] | https://github.com/soeaver/caffe-model | No |
+| resnet_v2_152 | tf1 | [1, 299, 299, 3] | https://github.com/tensorflow/models/tree/master/research/slim#Pretrained | No |
+| se_resnet_50 | tf1 | [1, 224, 224, 3] | https://github.com/HiKapok/TF-SENet | No |
+| srcnn | tf1 | [1, 33, 33, 1] | https://github.com/tegg89/SRCNN-Tensorflow | No |
+| nasnet_mobile | tflite | [1, 224, 224, 3] | https://www.tensorflow.org/lite/guide/hosted_models?hl=zh-cn | No |
+| super_resolution | onnx | [1, 1, 224, 224] | https://github.com/onnx/models/tree/master/vision/super_resolution/sub_pixel_cnn_2016 | No |
+| squeezenet | onnx | [1, 3, 224, 224] | https://github.com/onnx/models/tree/master/vision/classification/squeezenet | No |
+| resnet_v2_101 | caffe | [1, 3, 448, 448] | https://github.com/soeaver/caffe-model | No |
+| densenet_121 | caffe | [1, 3, 224, 224] | https://github.com/soeaver/caffe-model | No |
+| yolo_v2 | onnx | [1, 3, 416, 416] | https://github.com/onnx/models/tree/master/vision/object_detection_segmentation/yolov2-coco | No |
+| yolo_v2_tiny | onnx | [1, 3, 416, 416] | https://github.com/onnx/models/tree/master/vision/object_detection_segmentation/tiny-yolov2 | No |
+| inception_v2 | tf1 | [1, 224, 224, 3] | https://github.com/tensorflow/models/tree/master/research/slim#Pretrained | No |
+| lightface | onnx | [1, 3, 240, 320] | https://hailo.ai/devzone-model-zoo/face-detection/ | No |
+| face_boxes | onnx | [1,3,1024,1024] | https://github.com/zisianw/FaceBoxes.PyTorch | No |
+| vgg_cnn_s | caffe | [10, 3, 224, 224] | https://gist.github.com/ksimonyan/fd8800eeb36e276cd6f9 | No |
+| mobilenet_v1 | caffe | [1, 3, 224, 224] | https://github.com/shicai/MobileNet-Caffe | No |
+| regnet_x | onnx | [1, 3, 224, 224] | https://hailo.ai/devzone-model-zoo/about-object-detection/ | No |
+| yolact_regnetx | onnx | [1, 3, 512, 512] | https://hailo.ai/devzone-model-zoo/instance-segmentation/ | No |
+| inception_resnet_v2 | caffe | [1, 3,331,331] | https://github.com/soeaver/caffe-model | No |
+| resnext_50 | caffe | [1, 3, 224, 224] | https://github.com/soeaver/caffe-model | No |
+| resnet_v2_50 | tf1 | [1, 299, 299, 3] | https://github.com/tensorflow/models/tree/master/research/slim#Pretrained | No |
+| pnasnet_large | tf1 | [1, 331, 331, 3] | https://github.com/tensorflow/models/tree/master/research/slim#Pretrained | No |
+| fast_depth | onnx | [1,3,224,224] | https://github.com/dwofk/fast-depth | No |
+| efficientnet_lite | tf1 | [1, 280, 280, 3] | https://hailo.ai/devzone-model-zoo/about-object-detection/ | No |
+| centernet | tf1 | [1, 512, 512, 3] | https://hailo.ai/devzone-model-zoo/object-detection/ | No |
+| inception_v1 | onnx | [1, 3, 224, 224] | https://github.com/onnx/models/tree/master/vision/classification/inception_and_googlenet/inception_v1 | No |
+| unet_bio | tf1 | [1,256, 256,1] | https://github.com/zhixuhao/unet | No |
+| shufflenet_v1 | onnx | [1, 3, 224, 224] | https://github.com/onnx/models/tree/master/vision/classification/shufflenet | No |
+| resnext_101 | caffe | [1, 3, 224, 224] | https://www.deepdetect.com/models/resnext/ | No |
+| se_resnet_50 | caffe | [1, 3, 225, 225] | https://github.com/soeaver/caffe-model | No |
+| xception | caffe | [1, 3, 299, 299] | https://github.com/soeaver/caffe-model | No |
+| resnet_v1_152 | tf1 | [1, 224, 224, 3] | https://github.com/tensorflow/models/tree/master/research/slim#Pretrained | No |
+| densenet_169 | caffe | [1, 3, 224, 224] | https://github.com/soeaver/caffe-model | No |
+| mixnet | tf1 | [1,224,224,3] | https://github.com/openvinotoolkit/open_model_zoo/tree/master/models/public/mixnet-l | No |
+| mnasnet | tflite | [1, 224, 224, 3] | https://www.tensorflow.org/lite/guide/hosted_models?hl=zh-cn | No |
+| ann | onnx | [1, 3, 1024, 2048] | https://github.com/open-mmlab/mmsegmentation/tree/master/configs/ann | No |
+| duc | onnx | [1, 3, 800, 800] | https://github.com/onnx/models/tree/master/vision/object_detection_segmentation/duc | No |
+| efficientnet_lite | tflite | [1, 300, 300, 3] | https://tfhub.dev/tensorflow/efficientnet/lite4/classification/2 | No |
+| se_inception | caffe | [1, 3, 224, 224] | https://github.com/openvinotoolkit/open_model_zoo/blob/master/models/public/se-inception/README.md | No |
+| rnn_t_encoder | onnx | [1,249,240] | https://github.com/mlcommons/inference/tree/master/speech_recognition/rnnt | No |
+| nasnet_mobile | tf1 | [1, 224, 224, 3] | https://www.tensorflow.org/lite/guide/hosted_models?hl=zh-cn | No |
+| yolo_v5 | onnx | [1, 3, 640, 640] | https://github.com/ultralytics/yolov5 | No |
+| fcn16s | onnx | [1, 3, 224, 224] | https://github.com/wkentaro/pytorch-fcn/ | No |
+| dilation_8 | caffe | [1, 3, 900, 900] | https://github.com/fyu/dilation | No |
+| zfnet_512 | onnx | [1, 3, 224, 224] | https://github.com/onnx/models/tree/master/vision/classification/zfnet-512 | No |
+| mobilenet_v2_ssd_lite | tf1 | [1, 300, 300, 3] | https://github.com/tensorflow/models/blob/master/research/object_detection/g3doc/tf1_detection_zoo.md | No |
+| inception_v2 | tflite | [1, 224, 224, 3] | https://www.tensorflow.org/lite/guide/hosted_models?hl=zh-cn | No |
+| age_googlenet | onnx | [1, 3, 224, 224] | https://github.com/onnx/models/tree/master/vision/body_analysis/age_gender | No |
+| rnn_t_decoder | onnx | [1, 1],[1, 2, 320],[1, 2, 320],[1, 1, 1024] | https://github.com/mlcommons/inference/tree/master/speech_recognition/rnnt | No |
+| stacked_hourglass | tf1 | [1, 256, 256, 3] | https://github.com/yuanyuanli85/Stacked_Hourglass_Network_Keras | No |
+| resnet_v2_152 | caffe | [1, 3, 224, 224] | https://github.com/soeaver/caffe-model | No |
+| yolo_v1_tiny | caffe | [1, 3, 448, 448] | https://github.com/xingwangsfu/caffe-yolo | No |
+| googlenet | onnx | [1, 3, 224, 224] | https://github.com/onnx/models/tree/master/vision/classification/inception_and_googlenet/googlenet | No |
+| mobilenet_v1_224 | tf1 | [1, 224, 224, 3] | https://github.com/tensorflow/models/tree/master/research/slim#Pretrained | No |
+| poly_lanenet | onnx | [1, 3, 640, 640] | https://hailo.ai/devzone-model-zoo/driving-lane-detection/ | No |
+| squeezenet | caffe | [10, 3, 227, 227] | https://github.com/forresti/SqueezeNet | No |
+| inception_v2 | onnx | [1, 3, 224, 224] | https://github.com/onnx/models/tree/master/vision/classification/inception_and_googlenet/inception_v2 | No |
+| vgg_19 | tf1 | [1, 224, 224, 3] | https://github.com/tensorflow/models/tree/master/research/slim#Pretrained | No |
+| efficientnet_lite | onnx | [1,224,224,3] | https://github.com/onnx/models/tree/master/vision/classification/efficientnet-lite4 | No |
+| caffenet | caffe | [10, 3, 227, 227] | https://github.com/BVLC/caffe/tree/master/models/bvlc_reference_caffenet | No |
+| densenet | tflite | [1, 224, 224, 3] | https://www.tensorflow.org/lite/guide/hosted_models?hl=zh-cn | No |
+| mobilenet_v3 | tf1 | [1, 224, 224, 3] | https://github.com/tensorflow/models/tree/master/research/slim/nets/mobilenet | No |
+| googlenet | caffe | [10, 3, 224, 224] | https://github.com/BVLC/caffe/tree/master/models/bvlc_googlenet | No |
+| densenet_121 | onnx | [1, 3, 224, 224] | https://github.com/onnx/models/tree/master/vision/classification/densenet-121 | No |
+| nasnet_large | tf1 | [1, 331, 331,3] | https://www.tensorflow.org/lite/guide/hosted_models?hl=zh-cn | No |
+| arcface | onnx | [1, 3, 112, 112] | https://github.com/onnx/models/tree/master/vision/body_analysis/arcface | No |
+| dpn_68_extra | caffe | [1, 3, 224, 224] | https://github.com/soeaver/caffe-model | No |
+| bisenet_v2 | tf1 | [4, 512, 1024, 3] | https://github.com/MaybeShewill-CV/bisenetv2-tensorflow | No |
+| caffenet | onnx | [1, 3, 224, 224] | https://github.com/onnx/models/tree/master/vision/classification/caffenet | No |
+| yolo_v4 | onnx | [1, 416, 416, 3] | https://github.com/onnx/models/tree/master/vision/object_detection_segmentation/yolov4 | No |
+| inception_v1 | tf1 | [1, 224, 224, 3] | https://github.com/tensorflow/models/tree/master/research/slim#Pretrained | No |
+| fcn32s | onnx | [1, 3, 224, 224] | https://github.com/wkentaro/pytorch-fcn/ | No |
+| inception_v1 | tflite | [1, 224, 224, 3] | https://www.tensorflow.org/lite/guide/hosted_models?hl=zh-cn | No |
+| mobilenet_v1 | tflite | [1, 224, 224, 3] | https://www.tensorflow.org/lite/guide/hosted_models?hl=zh-cn | No |
+| resnet_v2_101 | tflite | [1, 299, 299, 3] | https://www.tensorflow.org/lite/guide/hosted_models?hl=zh-cn | No |
+| densenet_201 | caffe | [1, 3, 224, 224] | https://github.com/soeaver/caffe-model | No |
+| efficientnet_l | tf1 | [1, 300, 300, 3] | https://hailo.ai/devzone-model-zoo/about-object-detection/ | No |
+| pspnet | onnx | [1, 3, 1024, 2048] | https://github.com/open-mmlab/mmsegmentation/tree/master/configs/pspnet | No |
+| inception_resnet_v2 | tf1 | [1, 299, 299, 3] | https://github.com/tensorflow/models/tree/master/research/slim#Pretrained | No |
+| se_resnet_101 | caffe | [1, 3, 224, 224] | https://github.com/hujie-frank/SENet | No |
+| apcnet | onnx | [1, 3, 1024, 2048] | https://github.com/open-mmlab/mmsegmentation/tree/master/configs/apcnet | No |
+| inception_resnet_v2 | tflite | [1, 299, 299, 3] | https://www.tensorflow.org/lite/guide/hosted_models?hl=zh-cn | No |
+| efficientdet | tflite | [ 1,512, 512, 3] | https://tfhub.dev/tensorflow/lite-model/efficientdet/lite3/detection/default/1 | No |
+| yolo_v5_s | onnx | [1, 3, 640, 640] | https://github.com/ultralytics/yolov5/ | No |
+| gender_googlenet | onnx | [1, 3, 224, 224] | https://github.com/onnx/models/tree/master/vision/body_analysis/age_gender | No |
+| resnet_v1_ssd | tf1 | [1, 640, 640, 3] | https://github.com/tensorflow/models/blob/master/research/object_detection/g3doc/tf1_detection_zoo.md | No |
+| retinanet | onnx | [1, 3, 480, 640] | https://github.com/onnx/models/tree/master/vision/object_detection_segmentation/retinanet | No |
+| deeplab_v3_xception | tf1 | [1,513,513,3] | https://github.com/tensorflow/models/blob/66264b2353aeeca3d4b340908a9590571495d5a6/research/deeplab/g3doc/model_zoo.md | No |
+| yolo_v5 | tflite | [1, 640, 640, 3] | https://github.com/ultralytics/yolov5 | No |
+| deeplabv3_mobilenetv2 | onnx | [1, 3, 512, 512] | https://github.com/zym1119/DeepLabv3_MobileNetv2_PyTorch | No |
+| yolox_s | onnx | [1, 3, 640, 640] | https://github.com/Megvii-BaseDetection/YOLOX/tree/main/demo/ONNXRuntime | No |
+| mobilenet_v1_ssd | onnx | [1, 3, 416, 416] | https://github.com/onnx/models/tree/master/vision/object_detection_segmentation/ssd-mobilenetv1 | No |
+| vision_transformer | onnx | [1,3,224,224] | https://github.com/jeonsworld/ViT-pytorch | No |
+| ssd | onnx | [1, 3, 1200, 1200] | https://github.com/onnx/models/tree/master/vision/object_detection_segmentation/ssd | No |
+| inception_v2_ssd | tf1 | [1, 300, 300, 3] | https://github.com/tensorflow/models/blob/master/research/object_detection/g3doc/tf1_detection_zoo.md | No |
+| yolo_v4 | tf1 | [1,416,416,3] | https://github.com/hunglc007/tensorflow-yolov4-tflite | No |
+| mobilenet_v1_ssd | tf1 | [1, 300, 300, 3] | https://github.com/tensorflow/models/blob/master/research/object_detection/g3doc/tf1_detection_zoo.md | No |
+| fcn8s | onnx | [1, 3, 224, 224] | https://github.com/wkentaro/pytorch-fcn/ | No |
+| drn_38 | onnx | [1, 3, 512, 1024] | https://github.com/fyu/drn | No |
+| fsrcnn | onnx | [1, 1, 85, 85] | https://github.com/yjn870/FSRCNN-pytorch | No |
+| pointnet | onnx | [32, 9, 4096] | https://github.com/yanx27/Pointnet_Pointnet2_pytorch | No |
+| mobilenet_edgetpu | tflite | [1,224,224,3] | https://github.com/mlcommons/mobile_models/blob/main/v0_7/tflite/mobilenet_edgetpu_224_1.0_uint8.tflite | No |
+| gru_l | tf1 | [1, 49, 10] | https://github.com/UT2UH/ML-KWS-for-ESP32/tree/master/Pretrained_models/GRU | No |
+| swin_transformer | onnx | [1,3,384,384] | https://github.com/microsoft/Swin-Transformer | No |
+| codeformer_256 | onnx | [1,3,256,256] | https://github.com/sczhou/CodeFormer/releases/tag/v0.1.0 | No |
+| resnet_18 | onnx | [1, 3, 224, 224] | https://github.com/onnx/models/tree/main/vision/classification/resnet/model | No |
+| resnext_101 | onnx | [1, 3, 224, 224] | https://github.com/Cadene/pretrained-models.pytorch | No |
+| unet_3d | onnx | [1, 3, 224, 224, 32] | https://zenodo.org/record/3904138#.YbBtatDP1PY | No |
+| sne_roadseg | onnx | [1,3,384,1248] | https://github.com/hlwang1124/SNE-RoadSeg | No |
+| maskrcnn | pytorch | [0] | https://pytorch.org/vision/main/models/mask_rcnn.html | No |
+| yolo_v6s | onnx | [1,3,640,640] | https://github.com/DefTruth/lite.ai.toolkit | No |
+| swin_transformer_base | onnx | [1,3,224,224] | https://github.com/microsoft/Swin-Transformer/blob/main/MODELHUB.md | No |
+| 3d_unet | onnx | [0] | https://github.com/mlcommons/inference/tree/master/vision/medical_imaging/3d-unet-kits19 | No |
+| xception | tf2 | [1, 299, 299, 3] | https://www.tensorflow.org/versions/r2.6/api_docs/python/tf/keras/applications | No |
+| efficientnet_b5 | tf2 | [1,456,456,3] | https://www.tensorflow.org/versions/r2.6/api_docs/python/tf/keras/applications | No |
+| mobilenet_v1_ssd | tflite | [300, 300] | https://tfhub.dev/tensorflow/lite-model/ssd_mobilenet_v1/1/metadata/2 | No |
+| transformer_official | tf1 | [1, 32] | https://github.com/Kyubyong/transformer | No |
+| ViT_B_16 | pytorch | [1, 3, 224, 224] | https://pytorch.org/vision/stable/models/vision_transformer.html | No |
+| efficientnet_b4_quant | tflite | [1,380,380,3] | https://ai-benchmark.com/download.html | Yes |
+| dped_quant | tflite | [1,1536,2048,3] | https://ai-benchmark.com/download.html | Yes |
+| shufflenet_v2 | onnx | [1,3,224,224] | https://github.com/TropComplique/shufflenet-v2-tensorflow | Yes |
+| vgg_quant | tflite | [1,256,256,1] | https://ai-benchmark.com/download.html | Yes |
+| caffenet | onnx | [1, 3, 224, 224] | https://github.com/onnx/models/tree/master/vision/classification/caffenet | Yes |
+| mobilebert_quant | tflite | [1,384], [1,384], [1,384] | https://ai-benchmark.com/download.html | Yes |
+| shufflenet_v2 | pytorch | [1, 3, 224, 224] | https://pytorch.org/vision/stable/models/shufflenetv2.html | Yes |
+| fcn32s | onnx | [1, 3, 224, 224] | https://github.com/wkentaro/pytorch-fcn/ | Yes |
+| imdn_quant | tflite | [1,1024,1024,3] | https://ai-benchmark.com/download.html | Yes |
+| squeezenet | onnx | [1, 3, 224, 224] | https://github.com/onnx/models/tree/master/vision/classification/squeezenet | Yes |
+| resnet_v1_50 | pytorch | [1, 3, 224, 224] | https://pytorch.org/vision/stable/models/resnet.html | Yes |
+| vsr_quant | tflite | [1,540,960,3] | https://ai-benchmark.com/download.html | Yes |
+| punet_quant | tflite | [1,544,960,4] | https://ai-benchmark.com/download.html | Yes |
+| dped_instance_quant | tflite | [1,1024,1536,3] | https://ai-benchmark.com/download.html | Yes |
+| mobilenet_v2_quant | tflite | [1,224,224,3] | https://ai-benchmark.com/download.html | Yes |
+| vgg_16 | onnx | [1,3,224,224] | https://github.com/tensorflow/models/tree/archive/research/slim/nets | Yes |
+| mobilenet_v2_b8_quant | tflite | [8,224,224,3] | https://ai-benchmark.com/download.html | Yes |
+| unet_quant | tflite | [1,1024,1024,3] | https://ai-benchmark.com/download.html | Yes |
+| resnet_v1_50 | onnx | [1,3,224,224] | https://github.com/tensorflow/models/tree/archive/research/slim/nets | Yes |
+| mobilenet_v3_quant | tflite | [1,512,512,3] | https://ai-benchmark.com/download.html | Yes |
+| crnn_quant | tflite | [1,64,200,3] | https://ai-benchmark.com/download.html | Yes |
+| deeplab_v3_plus_quant | tflite | [1,1024,1024,3] | https://ai-benchmark.com/download.html | Yes |
+| age_googlenet | onnx | [1, 3, 224, 224] | https://github.com/onnx/models/tree/master/vision/body_analysis/age_gender | Yes |
+| googlenet | pytorch | [1, 3, 224, 224] | https://pytorch.org/vision/stable/models/googlenet.html | Yes |
+| inception_v3 | pytorch | [1,3,224,224] | https://pytorch.org/vision/stable/models/generated/torchvision.models.quantization.inception_v3.html#torchvision.models.quantization.inception_v3 | Yes |
+| srgan_quant | tflite | [1,256,256,3] | https://ai-benchmark.com/download.html | Yes |
+| mv3_depth_quant | tflite | [1,1024,1536,3] | https://ai-benchmark.com/download.html | Yes |
+| mobilenet_v3 | pytorch | [1, 3, 224, 224] | https://pytorch.org/vision/stable/models/mobilenetv3.html | Yes |
+| mobiledet_ssd | tflite | [1, 320, 320, 3] | https://github.com/tensorflow/models/blob/master/research/object_detection/g3doc/tf1_detection_zoo.md | Yes |
+| pynet_quant | tflite | [1,512,512,3] | https://ai-benchmark.com/download.html | Yes |
+| mobilenet_v2 | onnx | [1,3,224,224] | https://github.com/tensorflow/models/tree/archive/research/slim/nets/mobilenet | Yes |
+| lstm_quant | tflite | [1,32,500,1] | https://ai-benchmark.com/download.html | Yes |
+| yolo_v4_tiny_quant | tflite | [1,416,416,3] | https://ai-benchmark.com/download.html | Yes |
+| inception_v3_quant | tflite | [1, 346, 346, 3] | https://ai-benchmark.com/download.html | Yes |
+| esrgan_quant | tflite | [1,128,128,3] | https://ai-benchmark.com/download.html | Yes |
+| mobilenet_v3_b4_quant | tflite | [4,512,512,3] | https://ai-benchmark.com/download.html | Yes |
+| xlsr_quant | tflite | [1,360,640,3] | https://ai-benchmark.com/download.html | Yes |
+| efficientnet_b5 | pytorch | [1, 3, 456, 456] | https://pytorch.org/vision/stable/models/generated/torchvision.models.efficientnet_b5.html#torchvision.models.efficientnet_b5 | No |
+| petr | pytorch | [6, 3, 320, 800] | https://github.com/megvii-research/PETR | No |
+| fcn | pytorch | [1, 3, 224, 224] | https://pytorch.org/vision/stable/models/fcn.html | No |
+| centernet_resnet50 | pytorch | [1, 3, 512, 512] | https://github.com/bubbliiiing/centernet-pytorch | No |
+| resnet_34_ssd | tf1 | [1, 1200, 1200, 3] | https://github.com/mlcommons/inference/tree/r0.5/v0.5/classification_and_detection | No |
+| densenet_121 | tf2 | [1, 224, 224, 3] | https://www.tensorflow.org/versions/r2.6/api_docs/python/tf/keras/applications | No |
+| yolo_v5s6 | onnx | [1, 3, 640, 640] | https://github.com/ultralytics/yolov5/releases/download/v7.0/yolov5s6.pt | No |
+| resnet_v1_101 | tf2 | [1, 224, 224, 3] | https://www.tensorflow.org/versions/r2.6/api_docs/python/tf/keras/applications | No |
+| shufflenet_v2 | pytorch | [1,3,224,224] | https://pytorch.org/vision/stable/models/generated/torchvision.models.shufflenet_v2_x1_0.html#torchvision.models.shufflenet_v2_x1_0 | No |
+| yolo_v4 | pytorch | [1, 3, 416, 416] | https://github.com/bubbliiiing/yolov4-pytorch/tree/master | No |
+| resnet_50_v1_5 | tf1 | [1, 224, 224, 3] | https://zenodo.org/ | No |
+| alexnet | pytorch | [2,3,224,224] | https://pytorch.org/vision/stable/models/generated/torchvision.models.alexnet.html#torchvision.models.alexnet | No |
+| yolo_v4_tiny | pytorch | [1, 3, 416, 416] | https://github.com/bubbliiiing/yolov4-tiny-pytorch | No |
+| se_resnext_50 | tf1 | [1, 224, 224, 3] | https://github.com/HiKapok/TF-SENet | No |
+| vgg_16 | pytorch | [1, 3, 224, 224] | https://pytorch.org/vision/stable/models/vgg.html | No |
+| mobilenet_v2 | tf2 | [1, 224, 224, 3] | https://www.tensorflow.org/versions/r2.6/api_docs/python/tf/keras/applications | No |
+| sea_former | onnx | [1, 3, 512, 1024] | https://github.com/fudan-zvg/SeaFormer/tree/main/seaformer-seg | No |
+| resnet_v2_50 | tf2 | [1, 224, 224, 3] | https://www.tensorflow.org/versions/r2.6/api_docs/python/tf/keras/applications | No |
+| resnet_v2_101 | tf2 | [1, 224, 224, 3] | https://www.tensorflow.org/versions/r2.6/api_docs/python/tf/keras/applications | No |
+| vgg_16 | tf2 | [1, 224, 224, 3] | https://www.tensorflow.org/versions/r2.6/api_docs/python/tf/keras/applications | No |
+| mobilenet_v1_224 | tflite | [1, 224, 224, 3] | https://github.com/tensorflow/models/blob/master/research/slim/nets/mobilenet_v1.md | No |
+| vgg_19 | tf2 | [1, 224, 224, 3] | https://www.tensorflow.org/versions/r2.6/api_docs/python/tf/keras/applications | No |
+| yolox_l | onnx | [1,3,640,640] | https://github.com/Megvii-BaseDetection/YOLOX | No |
+| yolo_v4_tiny | tflite | [1,416,416,3] | https://github.com/hunglc007/tensorflow-yolov4-tflite/tree/master | No |
+| swin_transformer_tiny_224 | pytorch | [1, 3, 224, 224] | https://pytorch.org/vision/stable/models/swin_transformer.html | No |
+| inception_v3 | tf2 | [1, 299, 299, 3] | https://www.tensorflow.org/versions/r2.6/api_docs/python/tf/keras/applications | No |
+| nasnet_mobile | tf2 | [1, 224, 224, 3] | https://www.tensorflow.org/versions/r2.6/api_docs/python/tf/keras/applications | No |
+| mobilenet_v3 | tf2 | [1, 224, 224, 3] | https://www.tensorflow.org/versions/r2.6/api_docs/python/tf/keras/applications | No |
+| resnet_v1_50 | tf2 | [1, 224, 224, 3] | https://www.tensorflow.org/versions/r2.6/api_docs/python/tf/keras/applications | No |
+| resnet_v1_50 | pytorch | [1, 3, 224, 224] | https://pytorch.org/vision/stable/models/resnet.html | No |
diff --git a/docs/sirider/s1/app-development/zhouyi_npu.md b/docs/sirider/s1/app-development/zhouyi_npu.md
new file mode 100644
index 000000000..b06af0787
--- /dev/null
+++ b/docs/sirider/s1/app-development/zhouyi_npu.md
@@ -0,0 +1,148 @@
+---
+sidebar_position: 1
+---
+# 周易Z2 AIPU
+
+“周易” AIPU 是由安谋中国针对深度学习而自主研发的创新性 AI 专用处理器,它采用了创新性的架构设计,提供完整的硬件和软件生态,并且具有 PPA 最佳平衡。
+安谋中国还为“周易” AIPU 的客户提供很多工具来帮助他们进行开发,包括仿真器、编译器和调试器等进行数据的采集、分析。
+“周易” AIPU 也支持业界主流的 AI 规模框架,包括 TensorFlow、ONNX 等,未来也将支持更多不同的扩展框架。
+
+“周易” Z2 AIPU 将主要面向中高端安防、智能座舱和 ADAS、边缘服务器等应用场景。
+## 周易 Z2 AIPU 使用教程
+
+### x86 PC 端安装 周易 AIPU SDK
+周易 SDK 是一个全栈平台,可为用户提供快速上市的开发和部署能力。
+![image](https://user-images.githubusercontent.com/85479712/198521602-49e13a31-bb49-424f-b782-5108274d63c3.png)
+
+- 在[瑞莎下载站](https://dl.radxa.com/sirider/s1/)下载周易 Z2 SDK 安装包后解压安装
+ ```bash
+ tar -xvf Zhouyi_Z2.tar.gz
+ cd Zhouyi_Z2 && bash +x SETUP.SH
+ ```
+- 安装后得到的完整 SDK 文件如下
+
+ - `AI610-SDK-r1p3-AIoT` : ARM ZhouYi Z2 工具包
+
+ - `siengine` : siengine 提供的 ARM ZhouYi Z2 模型编译(nn-compiler-user-case-example)及板子部署(nn-runtime-user-case-example)的 demos
+
+- 配置 nn-compiler 环境
+ ```bash
+ cd AI610-SDK-r1p3-AIoT/AI610-SDK-r1p3-00eac0/Out-Of-Box/out-of-box-nn-compiler
+ pip3 install -r lib_dependency.txt
+ ```
+ 因为此 SDK 不包含模拟功能, 故安装过程会出现安装 AIPUSimProfiler 的报错,可以忽略
+
+ 若使用 venv 的用户请在 env_setup.sh 中 pip3 install 部分去掉 --user 选项
+ ```bash
+ source env_setup.sh
+ ```
+
+### x86 PC 端模型转换
+nn-compiler 可以将 TensorFlow、ONNX 等框架模型转换成可以在周易 AIPU 进行硬件加速推理的模型文件
+:::tip
+此案例中将介绍开箱即用案例:resnet50 目标分类
+
+完整 SDK 文档请参 `AI610-SDK-r1p3-AIoT/AI610-SDK-r1p3-00eac0/AI610-DOC-1001-r1p3-eac0`
+:::
+
+- 进入 siengine nn-compiler-user-case-example 目录
+
+ 如没配置好 nn-compiler 环境, 请按照 [x86 PC 端安装 AIPU SDK](#x86-pc-端安装-aipu-sdk)进行配置
+
+ ```bash
+ cd siengine/nn-compiler-user-case-example/onnx
+ ```
+
+- 生成量化校准集
+ ```bash
+ python3 generate_calibration_data.py
+ ```
+- 生成用于模型推理的照片文件
+ ```bash
+ python3 generate_input_binary.py
+ ```
+ 文件在 ./resnet50/input_3_224_224.bin
+
+- (可选) 配置 build.cfg (开箱即用案例已提供)
+ ```bash
+ vim ./resnet50/build.cfg
+ ```
+- 生成 aipu 模型
+ ```bash
+ cd ./restnet50
+ aipubuild build.cfg
+ ```
+ 在 ./restnet50 中得到 aipu_mlperf_resnet50.bin
+
+### 板端使用周易 Z2 推理 AIPU 模型
+在使用周易 Z2 AIPU 推理前需要在 x86 主机进行交叉编译生成可执行文件 `aiputest`,然后拷贝到 Sirider S1 中执行
+#### 在 x86 PC 端交叉编译二进制可执行文件
+- 安装 [gcc-linaro-7.5.0-2019.12-x86_64_aarch64-linux-gnu](https://releases.linaro.org/components/toolchain/binaries/latest-7/aarch64-linux-gnu/) 交叉编译工具链
+ ```bash
+ tar -xvf gcc-linaro-7.5.0-2019.12-x86_64_aarch64-linux-gnu.tar
+ cp -r gcc-linaro-7.5.0-2019.12-x86_64_aarch64-linux-gnu /opt
+ ```
+- 编译 aiputest
+
+ - 修改 UMDSRC 变量
+ ```bash
+ cd siengine/nn-runtime-user-case-example
+ vim CMakeLists.txt
+ #set(UMDSRC "${CMAKE_SOURCE_DIR}/../AI610-SDK-${AIPU_VERSION}-00eac0/AI610-SDK-1012-${AIPU_VERSION}-eac0/Linux-driver/driver/umd")
+ set(UMDSRC "${CMAKE_SOURCE_DIR}/../../AI610-SDK-${AIPU_VERSION}-AIoT/AI610-SDK-r1p3-00eac0/AI610-SDK-1012-${AIPU_VERSION}-eac0/Linux-driver/driver/umd")
+ ```
+ - 交叉编译
+ ```bash
+ mkdir build && cd build
+ cmake -DCMAKE_BUILD_TYPE=Release ..
+ make
+ ```
+ 编译生成的文件在 `siengine/nn-runtime-user-case-example/out/linux/aipu_test`
+
+#### 在 Sirider S1 进行板端推理
+- 将生成的 `aipu_mlperf_resnet50.bin` 模型文件,`input_3_224_224.bin` 照片文件,`aipu_test` 可执行文件,`out/linux/libs` 动态库文件夹复制到 Sirider S1 中
+- 执行 aipu_test
+ ```bash
+ export LD_LIBRARY_PATH=$LD_LIBRARY_PATH:
+ ./aipu_test aipu_mlperf_resnet50.bin input_3_224_224.bin
+ ```
+ ```bash
+ (aiot-focal_overlayfs)root@linux:~/ssd# ./aipu_test aipu_mlperf_resnet50.bin input_3_224_224.bin
+ usage: ./aipu_test aipu.bin input0.bin
+ aipu_init_context success
+ aipu_load_graph_helper success: aipu_mlperf_resnet50.bin
+ aipu_create_job success
+ Frame #0
+ aipu_finish_job success
+ No profiler data
+ get output tensor 0 success (1/1)
+ output_desc zero_point: 0.0000 scale: 5.5835
+ idx: 637 fval: 21.4919
+ idx: 749 fval: 19.8800
+ idx: 415 fval: 16.1189
+ idx: 412 fval: 15.0443
+ idx: 791 fval: 14.1488
+ Frame #1
+ aipu_finish_job success
+ No profiler data
+ get output tensor 0 success (1/1)
+ output_desc zero_point: 0.0000 scale: 5.5835
+ idx: 637 fval: 21.4919
+ idx: 749 fval: 19.8800
+ idx: 415 fval: 16.1189
+ idx: 412 fval: 15.0443
+ idx: 791 fval: 14.1488
+ aipu_clean_job success
+ aipu_unload_graph success
+ aipu_deinit_ctx success
+ ```
+ 两次的推理总时间
+ ```bash
+ real 0m0.043s
+ user 0m0.008s
+ sys 0m0.023s
+ ```
+
+ 这里结果仅显示 推理结果的标签值,最大置信度 637 即对应 [imagenet1000](https://gist.github.com/yrevar/942d3a0ac09ec9e5eb3a) 中的 `mailbag, postbag`
+
+ ![input.webp](/img/sirider/s1/aipu_1.webp)
\ No newline at end of file
diff --git a/i18n/en/docusaurus-plugin-content-docs/current/sirider/s1/app-development/README.md b/i18n/en/docusaurus-plugin-content-docs/current/sirider/s1/app-development/README.md
new file mode 100644
index 000000000..325ebf399
--- /dev/null
+++ b/i18n/en/docusaurus-plugin-content-docs/current/sirider/s1/app-development/README.md
@@ -0,0 +1,9 @@
+---
+sidebar_position: 4
+---
+
+# Application development
+
+Introduces upper-layer application development, such as QT, WiringX, Mraa, etc.
+
+
diff --git a/i18n/en/docusaurus-plugin-content-docs/current/sirider/s1/app-development/zhouyi_model_zoo.md b/i18n/en/docusaurus-plugin-content-docs/current/sirider/s1/app-development/zhouyi_model_zoo.md
new file mode 100644
index 000000000..93830a4b9
--- /dev/null
+++ b/i18n/en/docusaurus-plugin-content-docs/current/sirider/s1/app-development/zhouyi_model_zoo.md
@@ -0,0 +1,266 @@
+---
+sidebar_position: 2
+---
+
+# Zhouyi Model Zoo
+
+The [Zhouyi Model Zoo](https://github.com/Arm-China/Model_zoo) repository provides a set of AI models for reference used by Zhouyi SDK.
+
+#### **FTP Model Download (Recommended FTP Tool: [FileZilla](https://filezilla-project.org/))**
+ - `Host`: sftp://sftp01.armchina.com
+ - `Account`: zhouyi.armchina
+ - `Password`: 114r3cJd
+
+| Model | Framework | Input Shape | Model Source | Quant Model |
+|---------------------------|-----------|---------------------------------------------|---------------------------------------------------------------------------------------------------------------------------------------------------|-------------|
+| mobilenet_v2 | caffe | [1,3,224,224] | https://github.com/shicai/MobileNet-Caffe | No |
+| inception_v4 | tf1 | [1,299,299,3] | https://github.com/tensorflow/models/tree/archive/research/slim/nets | No |
+| deeplab_v2 | onnx | [1, 3, 513, 513] | https://github.com/kazuto1011/deeplab-pytorch | No |
+| mtcnn_o | caffe | [1,3,48,48] | https://github.com/CongWeilin/mtcnn-caffe | No |
+| inception_v4 | caffe | [1,3,299,299] | https://github.com/soeaver/caffe-model/tree/master/cls | No |
+| deepspeech_v2 | onnx | [1,385,161,1] | https://github.com/tensorflow/models/tree/archive/research/deep_speech | No |
+| wavenet | onnx | [1,390,23] | https://github.com/buriburisuri/speech-to-text-wavenet | No |
+| shufflenet_v2 | onnx | [1,3,224,224] | https://github.com/TropComplique/shufflenet-v2-tensorflow | No |
+| mobilenet_v2_ssd | onnx | [1,300,300,3] | https://github.com/tensorflow/models/tree/archive/research/object_detection/models | No |
+| facenet | tf1 | [1, 160, 160, 3] | https://github.com/davidsandberg/facenet | No |
+| yolo_v3_tiny | onnx | [1,416,416,3] | https://github.com/onnx/models/tree/main/vision/object_detection_segmentation/tiny-yolov3 | No |
+| yolo_v3 | caffe | [1, 3, 608, 608] | https://github.com/foss-for-synopsys-dwc-arc-processors/synopsys-caffe-models/tree/master/caffe_models/yolo_v3 | No |
+| yolo_v2_416 | tf1 | [1,416,416,3] | https://github.com/wojciechmo/yolo2 | No |
+| yolo_v2_416 | caffe | [1,3,416,416] | https://github.com/tsingjinyun/caffe-yolov2 | No |
+| yolo_v2_416 | onnx | [1,3,416,416] | https://github.com/wojciechmo/yolo2 | No |
+| inception_v4 | onnx | [1,3,299,299] | https://github.com/tensorflow/models/tree/archive/research/slim/nets | No |
+| mobilenet_v2 | onnx | [1,3,224,224] | https://github.com/tensorflow/models/tree/archive/research/slim/nets/mobilenet | No |
+| faster_rcnn | caffe | [1, 3, 224, 224] | https://github.com/rbgirshick/py-faster-rcnn | No |
+| shufflenet_v2 | tf1 | [1,224,224,3] | https://github.com/TropComplique/shufflenet-v2-tensorflow | No |
+| resnet_v1_101 | caffe | [1,3,224,224] | https://github.com/SnailTyan/caffe-model-zoo | No |
+| mtcnn_p | caffe | [1,3,12,12] | https://github.com/CongWeilin/mtcnn-caffe | No |
+| resnet_v1_50 | tf1 | [1,224,224,3] | https://github.com/tensorflow/models/tree/archive/research/slim/nets | No |
+| alexnet | onnx | [1,1,28,28] | https://github.com/tensorflow/models/tree/archive/research/slim/nets | No |
+| mobilenet_v2_ssd | tf1 | [1,300,300,3] | https://github.com/tensorflow/models/tree/archive/research/object_detection/models | No |
+| inception_v3 | tf1 | [1, 299, 299, 3] | https://github.com/tensorflow/models/tree/archive/research/slim/nets | No |
+| shufflenet_v2 | caffe | [1,3,224,224] | https://github.com/Ewenwan/ShuffleNet-2 | No |
+| deepspeech_v2 | tf1 | [1,385,161,1] | https://github.com/tensorflow/models/tree/archive/research/deep_speech | No |
+| inception_v3 | caffe | [1, 3, 299, 299] | https://github.com/soeaver/caffe-model/tree/master/cls | No |
+| mobilenet_v2_ssd | caffe | [1,3,300,300] | https://github.com/chuanqi305/MobileNet-SSD | No |
+| mtcnn_r | caffe | [1,3,24,24] | https://github.com/CongWeilin/mtcnn-caffe | No |
+| resnet_v1_101 | tf1 | [1,224,224,3] | https://github.com/tensorflow/models/tree/archive/research/slim/nets | No |
+| mobilenet_v2 | tf1 | [1,224,224,3] | https://github.com/tensorflow/models/tree/archive/research/slim/nets/mobilenet | No |
+| efficientnet_b5 | tf1 | [1, 456, 456, 3] | https://github.com/tensorflow/tpu/tree/master/models/official/efficientnet | No |
+| yolo_v3 | tf1 | [1, 416, 416, 3] | https://github.com/qqwweee/keras-yolo3 | No |
+| alexnet | tf1 | [1,28,28,1] | https://github.com/tensorflow/models/tree/archive/research/slim/nets | No |
+| resnet_v1_50 | onnx | [1,3,224,224] | https://github.com/tensorflow/models/tree/archive/research/slim/nets | No |
+| inception_v3 | onnx | [1, 3, 299, 299] | https://github.com/tensorflow/models/tree/archive/research/slim/nets | No |
+| vgg_16 | caffe | [1,3,224,224] | https://gist.github.com/ksimonyan/211839e770f7b538e2d8#file-readme-md | No |
+| resnet_v1_101 | onnx | [1,3,224,224] | https://github.com/tensorflow/models/tree/archive/research/slim/nets | No |
+| vgg_16 | onnx | [1,3,224,224] | https://github.com/tensorflow/models/tree/archive/research/slim/nets | No |
+| wavenet | tf1 | [1,390,23] | https://github.com/buriburisuri/speech-to-text-wavenet | No |
+| alexnet | tflite | [1,28,28,1] | https://github.com/tensorflow/models/tree/archive/research/slim/nets | No |
+| mobilenet_v2 | tflite | [1,224,224,3] | https://github.com/tensorflow/models/tree/archive/research/slim/nets/mobilenet | No |
+| shufflenet_v2 | tflite | [1,224,224,3] | https://github.com/TropComplique/shufflenet-v2-tensorflow | No |
+| inception_v3 | tflite | [1, 299, 299, 3] | https://github.com/tensorflow/models/tree/archive/research/slim/nets | No |
+| inception_v4 | tflite | [1,299,299,3] | https://github.com/tensorflow/models/tree/archive/research/slim/nets | No |
+| resnet_v1_50 | tflite | [1,224,224,3] | https://github.com/tensorflow/models/tree/archive/research/slim/nets | No |
+| resnet_v1_101 | tflite | [1,224,224,3] | https://github.com/tensorflow/models/tree/archive/research/slim/nets | No |
+| vgg_16 | tflite | [1,224,224,3] | https://github.com/tensorflow/models/tree/archive/research/slim/nets | No |
+| alexnet | caffe | [1,3,227,227] | https://github.com/BVLC/caffe/tree/master/models/bvlc_alexnet | No |
+| vgg_16 | tf1 | [1,224,224,3] | https://github.com/tensorflow/models/tree/archive/research/slim/nets | No |
+| resnet_v1_50 | caffe | [1,3,224,224] | https://github.com/SnailTyan/caffe-model-zoo | No |
+| deeplab_v3 | tflite | [1,257,257,3] | https://github.com/tensorflow/models/tree/archive/research/deeplab | No |
+| mobilenet_v3 | tflite | [1,224,224,3] | https://github.com/tensorflow/models/tree/master/research/slim/nets/mobilenet | No |
+| peleenet | caffe | [1,3,224,224] | https://github.com/Robert-JunWang/PeleeNet/tree/master/caffe | No |
+| fcn | caffe | [1,3,224,224] | https://github.com/shelhamer/fcn.berkeleyvision.org/tree/master/voc-fcn8s-atonce | No |
+| deeplab_v3 | onnx | [1,3,513,513] | https://github.com/tensorflow/models/tree/archive/research/deeplab | No |
+| erfnet | caffe | [1,3,512,1024] | https://github.com/Yuelong-Yu/ERFNet-Caffe | No |
+| deeplab_v3 | tf1 | [1,513,513,3] | https://github.com/tensorflow/models/tree/archive/research/deeplab | No |
+| centerface | onnx | [1, 3, 640, 640] | https://github.com/ttanzhiqiang/onnx_tensorrt_project | No |
+| vgg_ssd | caffe | [1, 3, 300, 300] | https://github.com/weiliu89/caffe/tree/ssd | No |
+| icnet | caffe | [1, 3, 1025, 2049] | https://github.com/hszhao/ICNet | No |
+| resnet_34_ssd | onnx | [1, 3, 1200, 1200] | https://github.com/mlcommons/inference/tree/master/vision/classification_and_detection | No |
+| resnet_v2_101 | tf1 | [1, 299, 299, 3] | https://github.com/tensorflow/models/tree/master/research/slim#Pretrained | No |
+| squeezenet | tflite | [1, 224, 224, 3] | https://www.tensorflow.org/lite/guide/hosted_models?hl=zh-cn | No |
+| sphereface | caffe | [1, 3, 112, 96] | https://github.com/openvinotoolkit/open_model_zoo/blob/master/models/public/Sphereface/README.md | No |
+| dpn_92 | caffe | [1, 3, 224, 224] | https://github.com/soeaver/caffe-model | No |
+| resnet_v2_152 | tf1 | [1, 299, 299, 3] | https://github.com/tensorflow/models/tree/master/research/slim#Pretrained | No |
+| se_resnet_50 | tf1 | [1, 224, 224, 3] | https://github.com/HiKapok/TF-SENet | No |
+| srcnn | tf1 | [1, 33, 33, 1] | https://github.com/tegg89/SRCNN-Tensorflow | No |
+| nasnet_mobile | tflite | [1, 224, 224, 3] | https://www.tensorflow.org/lite/guide/hosted_models?hl=zh-cn | No |
+| super_resolution | onnx | [1, 1, 224, 224] | https://github.com/onnx/models/tree/master/vision/super_resolution/sub_pixel_cnn_2016 | No |
+| squeezenet | onnx | [1, 3, 224, 224] | https://github.com/onnx/models/tree/master/vision/classification/squeezenet | No |
+| resnet_v2_101 | caffe | [1, 3, 448, 448] | https://github.com/soeaver/caffe-model | No |
+| densenet_121 | caffe | [1, 3, 224, 224] | https://github.com/soeaver/caffe-model | No |
+| yolo_v2 | onnx | [1, 3, 416, 416] | https://github.com/onnx/models/tree/master/vision/object_detection_segmentation/yolov2-coco | No |
+| yolo_v2_tiny | onnx | [1, 3, 416, 416] | https://github.com/onnx/models/tree/master/vision/object_detection_segmentation/tiny-yolov2 | No |
+| inception_v2 | tf1 | [1, 224, 224, 3] | https://github.com/tensorflow/models/tree/master/research/slim#Pretrained | No |
+| lightface | onnx | [1, 3, 240, 320] | https://hailo.ai/devzone-model-zoo/face-detection/ | No |
+| face_boxes | onnx | [1,3,1024,1024] | https://github.com/zisianw/FaceBoxes.PyTorch | No |
+| vgg_cnn_s | caffe | [10, 3, 224, 224] | https://gist.github.com/ksimonyan/fd8800eeb36e276cd6f9 | No |
+| mobilenet_v1 | caffe | [1, 3, 224, 224] | https://github.com/shicai/MobileNet-Caffe | No |
+| regnet_x | onnx | [1, 3, 224, 224] | https://hailo.ai/devzone-model-zoo/about-object-detection/ | No |
+| yolact_regnetx | onnx | [1, 3, 512, 512] | https://hailo.ai/devzone-model-zoo/instance-segmentation/ | No |
+| inception_resnet_v2 | caffe | [1, 3,331,331] | https://github.com/soeaver/caffe-model | No |
+| resnext_50 | caffe | [1, 3, 224, 224] | https://github.com/soeaver/caffe-model | No |
+| resnet_v2_50 | tf1 | [1, 299, 299, 3] | https://github.com/tensorflow/models/tree/master/research/slim#Pretrained | No |
+| pnasnet_large | tf1 | [1, 331, 331, 3] | https://github.com/tensorflow/models/tree/master/research/slim#Pretrained | No |
+| fast_depth | onnx | [1,3,224,224] | https://github.com/dwofk/fast-depth | No |
+| efficientnet_lite | tf1 | [1, 280, 280, 3] | https://hailo.ai/devzone-model-zoo/about-object-detection/ | No |
+| centernet | tf1 | [1, 512, 512, 3] | https://hailo.ai/devzone-model-zoo/object-detection/ | No |
+| inception_v1 | onnx | [1, 3, 224, 224] | https://github.com/onnx/models/tree/master/vision/classification/inception_and_googlenet/inception_v1 | No |
+| unet_bio | tf1 | [1,256, 256,1] | https://github.com/zhixuhao/unet | No |
+| shufflenet_v1 | onnx | [1, 3, 224, 224] | https://github.com/onnx/models/tree/master/vision/classification/shufflenet | No |
+| resnext_101 | caffe | [1, 3, 224, 224] | https://www.deepdetect.com/models/resnext/ | No |
+| se_resnet_50 | caffe | [1, 3, 225, 225] | https://github.com/soeaver/caffe-model | No |
+| xception | caffe | [1, 3, 299, 299] | https://github.com/soeaver/caffe-model | No |
+| resnet_v1_152 | tf1 | [1, 224, 224, 3] | https://github.com/tensorflow/models/tree/master/research/slim#Pretrained | No |
+| densenet_169 | caffe | [1, 3, 224, 224] | https://github.com/soeaver/caffe-model | No |
+| mixnet | tf1 | [1,224,224,3] | https://github.com/openvinotoolkit/open_model_zoo/tree/master/models/public/mixnet-l | No |
+| mnasnet | tflite | [1, 224, 224, 3] | https://www.tensorflow.org/lite/guide/hosted_models?hl=zh-cn | No |
+| ann | onnx | [1, 3, 1024, 2048] | https://github.com/open-mmlab/mmsegmentation/tree/master/configs/ann | No |
+| duc | onnx | [1, 3, 800, 800] | https://github.com/onnx/models/tree/master/vision/object_detection_segmentation/duc | No |
+| efficientnet_lite | tflite | [1, 300, 300, 3] | https://tfhub.dev/tensorflow/efficientnet/lite4/classification/2 | No |
+| se_inception | caffe | [1, 3, 224, 224] | https://github.com/openvinotoolkit/open_model_zoo/blob/master/models/public/se-inception/README.md | No |
+| rnn_t_encoder | onnx | [1,249,240] | https://github.com/mlcommons/inference/tree/master/speech_recognition/rnnt | No |
+| nasnet_mobile | tf1 | [1, 224, 224, 3] | https://www.tensorflow.org/lite/guide/hosted_models?hl=zh-cn | No |
+| yolo_v5 | onnx | [1, 3, 640, 640] | https://github.com/ultralytics/yolov5 | No |
+| fcn16s | onnx | [1, 3, 224, 224] | https://github.com/wkentaro/pytorch-fcn/ | No |
+| dilation_8 | caffe | [1, 3, 900, 900] | https://github.com/fyu/dilation | No |
+| zfnet_512 | onnx | [1, 3, 224, 224] | https://github.com/onnx/models/tree/master/vision/classification/zfnet-512 | No |
+| mobilenet_v2_ssd_lite | tf1 | [1, 300, 300, 3] | https://github.com/tensorflow/models/blob/master/research/object_detection/g3doc/tf1_detection_zoo.md | No |
+| inception_v2 | tflite | [1, 224, 224, 3] | https://www.tensorflow.org/lite/guide/hosted_models?hl=zh-cn | No |
+| age_googlenet | onnx | [1, 3, 224, 224] | https://github.com/onnx/models/tree/master/vision/body_analysis/age_gender | No |
+| rnn_t_decoder | onnx | [1, 1],[1, 2, 320],[1, 2, 320],[1, 1, 1024] | https://github.com/mlcommons/inference/tree/master/speech_recognition/rnnt | No |
+| stacked_hourglass | tf1 | [1, 256, 256, 3] | https://github.com/yuanyuanli85/Stacked_Hourglass_Network_Keras | No |
+| resnet_v2_152 | caffe | [1, 3, 224, 224] | https://github.com/soeaver/caffe-model | No |
+| yolo_v1_tiny | caffe | [1, 3, 448, 448] | https://github.com/xingwangsfu/caffe-yolo | No |
+| googlenet | onnx | [1, 3, 224, 224] | https://github.com/onnx/models/tree/master/vision/classification/inception_and_googlenet/googlenet | No |
+| mobilenet_v1_224 | tf1 | [1, 224, 224, 3] | https://github.com/tensorflow/models/tree/master/research/slim#Pretrained | No |
+| poly_lanenet | onnx | [1, 3, 640, 640] | https://hailo.ai/devzone-model-zoo/driving-lane-detection/ | No |
+| squeezenet | caffe | [10, 3, 227, 227] | https://github.com/forresti/SqueezeNet | No |
+| inception_v2 | onnx | [1, 3, 224, 224] | https://github.com/onnx/models/tree/master/vision/classification/inception_and_googlenet/inception_v2 | No |
+| vgg_19 | tf1 | [1, 224, 224, 3] | https://github.com/tensorflow/models/tree/master/research/slim#Pretrained | No |
+| efficientnet_lite | onnx | [1,224,224,3] | https://github.com/onnx/models/tree/master/vision/classification/efficientnet-lite4 | No |
+| caffenet | caffe | [10, 3, 227, 227] | https://github.com/BVLC/caffe/tree/master/models/bvlc_reference_caffenet | No |
+| densenet | tflite | [1, 224, 224, 3] | https://www.tensorflow.org/lite/guide/hosted_models?hl=zh-cn | No |
+| mobilenet_v3 | tf1 | [1, 224, 224, 3] | https://github.com/tensorflow/models/tree/master/research/slim/nets/mobilenet | No |
+| googlenet | caffe | [10, 3, 224, 224] | https://github.com/BVLC/caffe/tree/master/models/bvlc_googlenet | No |
+| densenet_121 | onnx | [1, 3, 224, 224] | https://github.com/onnx/models/tree/master/vision/classification/densenet-121 | No |
+| nasnet_large | tf1 | [1, 331, 331,3] | https://www.tensorflow.org/lite/guide/hosted_models?hl=zh-cn | No |
+| arcface | onnx | [1, 3, 112, 112] | https://github.com/onnx/models/tree/master/vision/body_analysis/arcface | No |
+| dpn_68_extra | caffe | [1, 3, 224, 224] | https://github.com/soeaver/caffe-model | No |
+| bisenet_v2 | tf1 | [4, 512, 1024, 3] | https://github.com/MaybeShewill-CV/bisenetv2-tensorflow | No |
+| caffenet | onnx | [1, 3, 224, 224] | https://github.com/onnx/models/tree/master/vision/classification/caffenet | No |
+| yolo_v4 | onnx | [1, 416, 416, 3] | https://github.com/onnx/models/tree/master/vision/object_detection_segmentation/yolov4 | No |
+| inception_v1 | tf1 | [1, 224, 224, 3] | https://github.com/tensorflow/models/tree/master/research/slim#Pretrained | No |
+| fcn32s | onnx | [1, 3, 224, 224] | https://github.com/wkentaro/pytorch-fcn/ | No |
+| inception_v1 | tflite | [1, 224, 224, 3] | https://www.tensorflow.org/lite/guide/hosted_models?hl=zh-cn | No |
+| mobilenet_v1 | tflite | [1, 224, 224, 3] | https://www.tensorflow.org/lite/guide/hosted_models?hl=zh-cn | No |
+| resnet_v2_101 | tflite | [1, 299, 299, 3] | https://www.tensorflow.org/lite/guide/hosted_models?hl=zh-cn | No |
+| densenet_201 | caffe | [1, 3, 224, 224] | https://github.com/soeaver/caffe-model | No |
+| efficientnet_l | tf1 | [1, 300, 300, 3] | https://hailo.ai/devzone-model-zoo/about-object-detection/ | No |
+| pspnet | onnx | [1, 3, 1024, 2048] | https://github.com/open-mmlab/mmsegmentation/tree/master/configs/pspnet | No |
+| inception_resnet_v2 | tf1 | [1, 299, 299, 3] | https://github.com/tensorflow/models/tree/master/research/slim#Pretrained | No |
+| se_resnet_101 | caffe | [1, 3, 224, 224] | https://github.com/hujie-frank/SENet | No |
+| apcnet | onnx | [1, 3, 1024, 2048] | https://github.com/open-mmlab/mmsegmentation/tree/master/configs/apcnet | No |
+| inception_resnet_v2 | tflite | [1, 299, 299, 3] | https://www.tensorflow.org/lite/guide/hosted_models?hl=zh-cn | No |
+| efficientdet | tflite | [ 1,512, 512, 3] | https://tfhub.dev/tensorflow/lite-model/efficientdet/lite3/detection/default/1 | No |
+| yolo_v5_s | onnx | [1, 3, 640, 640] | https://github.com/ultralytics/yolov5/ | No |
+| gender_googlenet | onnx | [1, 3, 224, 224] | https://github.com/onnx/models/tree/master/vision/body_analysis/age_gender | No |
+| resnet_v1_ssd | tf1 | [1, 640, 640, 3] | https://github.com/tensorflow/models/blob/master/research/object_detection/g3doc/tf1_detection_zoo.md | No |
+| retinanet | onnx | [1, 3, 480, 640] | https://github.com/onnx/models/tree/master/vision/object_detection_segmentation/retinanet | No |
+| deeplab_v3_xception | tf1 | [1,513,513,3] | https://github.com/tensorflow/models/blob/66264b2353aeeca3d4b340908a9590571495d5a6/research/deeplab/g3doc/model_zoo.md | No |
+| yolo_v5 | tflite | [1, 640, 640, 3] | https://github.com/ultralytics/yolov5 | No |
+| deeplabv3_mobilenetv2 | onnx | [1, 3, 512, 512] | https://github.com/zym1119/DeepLabv3_MobileNetv2_PyTorch | No |
+| yolox_s | onnx | [1, 3, 640, 640] | https://github.com/Megvii-BaseDetection/YOLOX/tree/main/demo/ONNXRuntime | No |
+| mobilenet_v1_ssd | onnx | [1, 3, 416, 416] | https://github.com/onnx/models/tree/master/vision/object_detection_segmentation/ssd-mobilenetv1 | No |
+| vision_transformer | onnx | [1,3,224,224] | https://github.com/jeonsworld/ViT-pytorch | No |
+| ssd | onnx | [1, 3, 1200, 1200] | https://github.com/onnx/models/tree/master/vision/object_detection_segmentation/ssd | No |
+| inception_v2_ssd | tf1 | [1, 300, 300, 3] | https://github.com/tensorflow/models/blob/master/research/object_detection/g3doc/tf1_detection_zoo.md | No |
+| yolo_v4 | tf1 | [1,416,416,3] | https://github.com/hunglc007/tensorflow-yolov4-tflite | No |
+| mobilenet_v1_ssd | tf1 | [1, 300, 300, 3] | https://github.com/tensorflow/models/blob/master/research/object_detection/g3doc/tf1_detection_zoo.md | No |
+| fcn8s | onnx | [1, 3, 224, 224] | https://github.com/wkentaro/pytorch-fcn/ | No |
+| drn_38 | onnx | [1, 3, 512, 1024] | https://github.com/fyu/drn | No |
+| fsrcnn | onnx | [1, 1, 85, 85] | https://github.com/yjn870/FSRCNN-pytorch | No |
+| pointnet | onnx | [32, 9, 4096] | https://github.com/yanx27/Pointnet_Pointnet2_pytorch | No |
+| mobilenet_edgetpu | tflite | [1,224,224,3] | https://github.com/mlcommons/mobile_models/blob/main/v0_7/tflite/mobilenet_edgetpu_224_1.0_uint8.tflite | No |
+| gru_l | tf1 | [1, 49, 10] | https://github.com/UT2UH/ML-KWS-for-ESP32/tree/master/Pretrained_models/GRU | No |
+| swin_transformer | onnx | [1,3,384,384] | https://github.com/microsoft/Swin-Transformer | No |
+| codeformer_256 | onnx | [1,3,256,256] | https://github.com/sczhou/CodeFormer/releases/tag/v0.1.0 | No |
+| resnet_18 | onnx | [1, 3, 224, 224] | https://github.com/onnx/models/tree/main/vision/classification/resnet/model | No |
+| resnext_101 | onnx | [1, 3, 224, 224] | https://github.com/Cadene/pretrained-models.pytorch | No |
+| unet_3d | onnx | [1, 3, 224, 224, 32] | https://zenodo.org/record/3904138#.YbBtatDP1PY | No |
+| sne_roadseg | onnx | [1,3,384,1248] | https://github.com/hlwang1124/SNE-RoadSeg | No |
+| maskrcnn | pytorch | [0] | https://pytorch.org/vision/main/models/mask_rcnn.html | No |
+| yolo_v6s | onnx | [1,3,640,640] | https://github.com/DefTruth/lite.ai.toolkit | No |
+| swin_transformer_base | onnx | [1,3,224,224] | https://github.com/microsoft/Swin-Transformer/blob/main/MODELHUB.md | No |
+| 3d_unet | onnx | [0] | https://github.com/mlcommons/inference/tree/master/vision/medical_imaging/3d-unet-kits19 | No |
+| xception | tf2 | [1, 299, 299, 3] | https://www.tensorflow.org/versions/r2.6/api_docs/python/tf/keras/applications | No |
+| efficientnet_b5 | tf2 | [1,456,456,3] | https://www.tensorflow.org/versions/r2.6/api_docs/python/tf/keras/applications | No |
+| mobilenet_v1_ssd | tflite | [300, 300] | https://tfhub.dev/tensorflow/lite-model/ssd_mobilenet_v1/1/metadata/2 | No |
+| transformer_official | tf1 | [1, 32] | https://github.com/Kyubyong/transformer | No |
+| ViT_B_16 | pytorch | [1, 3, 224, 224] | https://pytorch.org/vision/stable/models/vision_transformer.html | No |
+| efficientnet_b4_quant | tflite | [1,380,380,3] | https://ai-benchmark.com/download.html | Yes |
+| dped_quant | tflite | [1,1536,2048,3] | https://ai-benchmark.com/download.html | Yes |
+| shufflenet_v2 | onnx | [1,3,224,224] | https://github.com/TropComplique/shufflenet-v2-tensorflow | Yes |
+| vgg_quant | tflite | [1,256,256,1] | https://ai-benchmark.com/download.html | Yes |
+| caffenet | onnx | [1, 3, 224, 224] | https://github.com/onnx/models/tree/master/vision/classification/caffenet | Yes |
+| mobilebert_quant | tflite | [1,384], [1,384], [1,384] | https://ai-benchmark.com/download.html | Yes |
+| shufflenet_v2 | pytorch | [1, 3, 224, 224] | https://pytorch.org/vision/stable/models/shufflenetv2.html | Yes |
+| fcn32s | onnx | [1, 3, 224, 224] | https://github.com/wkentaro/pytorch-fcn/ | Yes |
+| imdn_quant | tflite | [1,1024,1024,3] | https://ai-benchmark.com/download.html | Yes |
+| squeezenet | onnx | [1, 3, 224, 224] | https://github.com/onnx/models/tree/master/vision/classification/squeezenet | Yes |
+| resnet_v1_50 | pytorch | [1, 3, 224, 224] | https://pytorch.org/vision/stable/models/resnet.html | Yes |
+| vsr_quant | tflite | [1,540,960,3] | https://ai-benchmark.com/download.html | Yes |
+| punet_quant | tflite | [1,544,960,4] | https://ai-benchmark.com/download.html | Yes |
+| dped_instance_quant | tflite | [1,1024,1536,3] | https://ai-benchmark.com/download.html | Yes |
+| mobilenet_v2_quant | tflite | [1,224,224,3] | https://ai-benchmark.com/download.html | Yes |
+| vgg_16 | onnx | [1,3,224,224] | https://github.com/tensorflow/models/tree/archive/research/slim/nets | Yes |
+| mobilenet_v2_b8_quant | tflite | [8,224,224,3] | https://ai-benchmark.com/download.html | Yes |
+| unet_quant | tflite | [1,1024,1024,3] | https://ai-benchmark.com/download.html | Yes |
+| resnet_v1_50 | onnx | [1,3,224,224] | https://github.com/tensorflow/models/tree/archive/research/slim/nets | Yes |
+| mobilenet_v3_quant | tflite | [1,512,512,3] | https://ai-benchmark.com/download.html | Yes |
+| crnn_quant | tflite | [1,64,200,3] | https://ai-benchmark.com/download.html | Yes |
+| deeplab_v3_plus_quant | tflite | [1,1024,1024,3] | https://ai-benchmark.com/download.html | Yes |
+| age_googlenet | onnx | [1, 3, 224, 224] | https://github.com/onnx/models/tree/master/vision/body_analysis/age_gender | Yes |
+| googlenet | pytorch | [1, 3, 224, 224] | https://pytorch.org/vision/stable/models/googlenet.html | Yes |
+| inception_v3 | pytorch | [1,3,224,224] | https://pytorch.org/vision/stable/models/generated/torchvision.models.quantization.inception_v3.html#torchvision.models.quantization.inception_v3 | Yes |
+| srgan_quant | tflite | [1,256,256,3] | https://ai-benchmark.com/download.html | Yes |
+| mv3_depth_quant | tflite | [1,1024,1536,3] | https://ai-benchmark.com/download.html | Yes |
+| mobilenet_v3 | pytorch | [1, 3, 224, 224] | https://pytorch.org/vision/stable/models/mobilenetv3.html | Yes |
+| mobiledet_ssd | tflite | [1, 320, 320, 3] | https://github.com/tensorflow/models/blob/master/research/object_detection/g3doc/tf1_detection_zoo.md | Yes |
+| pynet_quant | tflite | [1,512,512,3] | https://ai-benchmark.com/download.html | Yes |
+| mobilenet_v2 | onnx | [1,3,224,224] | https://github.com/tensorflow/models/tree/archive/research/slim/nets/mobilenet | Yes |
+| lstm_quant | tflite | [1,32,500,1] | https://ai-benchmark.com/download.html | Yes |
+| yolo_v4_tiny_quant | tflite | [1,416,416,3] | https://ai-benchmark.com/download.html | Yes |
+| inception_v3_quant | tflite | [1, 346, 346, 3] | https://ai-benchmark.com/download.html | Yes |
+| esrgan_quant | tflite | [1,128,128,3] | https://ai-benchmark.com/download.html | Yes |
+| mobilenet_v3_b4_quant | tflite | [4,512,512,3] | https://ai-benchmark.com/download.html | Yes |
+| xlsr_quant | tflite | [1,360,640,3] | https://ai-benchmark.com/download.html | Yes |
+| efficientnet_b5 | pytorch | [1, 3, 456, 456] | https://pytorch.org/vision/stable/models/generated/torchvision.models.efficientnet_b5.html#torchvision.models.efficientnet_b5 | No |
+| petr | pytorch | [6, 3, 320, 800] | https://github.com/megvii-research/PETR | No |
+| fcn | pytorch | [1, 3, 224, 224] | https://pytorch.org/vision/stable/models/fcn.html | No |
+| centernet_resnet50 | pytorch | [1, 3, 512, 512] | https://github.com/bubbliiiing/centernet-pytorch | No |
+| resnet_34_ssd | tf1 | [1, 1200, 1200, 3] | https://github.com/mlcommons/inference/tree/r0.5/v0.5/classification_and_detection | No |
+| densenet_121 | tf2 | [1, 224, 224, 3] | https://www.tensorflow.org/versions/r2.6/api_docs/python/tf/keras/applications | No |
+| yolo_v5s6 | onnx | [1, 3, 640, 640] | https://github.com/ultralytics/yolov5/releases/download/v7.0/yolov5s6.pt | No |
+| resnet_v1_101 | tf2 | [1, 224, 224, 3] | https://www.tensorflow.org/versions/r2.6/api_docs/python/tf/keras/applications | No |
+| shufflenet_v2 | pytorch | [1,3,224,224] | https://pytorch.org/vision/stable/models/generated/torchvision.models.shufflenet_v2_x1_0.html#torchvision.models.shufflenet_v2_x1_0 | No |
+| yolo_v4 | pytorch | [1, 3, 416, 416] | https://github.com/bubbliiiing/yolov4-pytorch/tree/master | No |
+| resnet_50_v1_5 | tf1 | [1, 224, 224, 3] | https://zenodo.org/ | No |
+| alexnet | pytorch | [2,3,224,224] | https://pytorch.org/vision/stable/models/generated/torchvision.models.alexnet.html#torchvision.models.alexnet | No |
+| yolo_v4_tiny | pytorch | [1, 3, 416, 416] | https://github.com/bubbliiiing/yolov4-tiny-pytorch | No |
+| se_resnext_50 | tf1 | [1, 224, 224, 3] | https://github.com/HiKapok/TF-SENet | No |
+| vgg_16 | pytorch | [1, 3, 224, 224] | https://pytorch.org/vision/stable/models/vgg.html | No |
+| mobilenet_v2 | tf2 | [1, 224, 224, 3] | https://www.tensorflow.org/versions/r2.6/api_docs/python/tf/keras/applications | No |
+| sea_former | onnx | [1, 3, 512, 1024] | https://github.com/fudan-zvg/SeaFormer/tree/main/seaformer-seg | No |
+| resnet_v2_50 | tf2 | [1, 224, 224, 3] | https://www.tensorflow.org/versions/r2.6/api_docs/python/tf/keras/applications | No |
+| resnet_v2_101 | tf2 | [1, 224, 224, 3] | https://www.tensorflow.org/versions/r2.6/api_docs/python/tf/keras/applications | No |
+| vgg_16 | tf2 | [1, 224, 224, 3] | https://www.tensorflow.org/versions/r2.6/api_docs/python/tf/keras/applications | No |
+| mobilenet_v1_224 | tflite | [1, 224, 224, 3] | https://github.com/tensorflow/models/blob/master/research/slim/nets/mobilenet_v1.md | No |
+| vgg_19 | tf2 | [1, 224, 224, 3] | https://www.tensorflow.org/versions/r2.6/api_docs/python/tf/keras/applications | No |
+| yolox_l | onnx | [1,3,640,640] | https://github.com/Megvii-BaseDetection/YOLOX | No |
+| yolo_v4_tiny | tflite | [1,416,416,3] | https://github.com/hunglc007/tensorflow-yolov4-tflite/tree/master | No |
+| swin_transformer_tiny_224 | pytorch | [1, 3, 224, 224] | https://pytorch.org/vision/stable/models/swin_transformer.html | No |
+| inception_v3 | tf2 | [1, 299, 299, 3] | https://www.tensorflow.org/versions/r2.6/api_docs/python/tf/keras/applications | No |
+| nasnet_mobile | tf2 | [1, 224, 224, 3] | https://www.tensorflow.org/versions/r2.6/api_docs/python/tf/keras/applications | No |
+| mobilenet_v3 | tf2 | [1, 224, 224, 3] | https://www.tensorflow.org/versions/r2.6/api_docs/python/tf/keras/applications | No |
+| resnet_v1_50 | tf2 | [1, 224, 224, 3] | https://www.tensorflow.org/versions/r2.6/api_docs/python/tf/keras/applications | No |
+| resnet_v1_50 | pytorch | [1, 3, 224, 224] | https://pytorch.org/vision/stable/models/resnet.html | No |
diff --git a/i18n/en/docusaurus-plugin-content-docs/current/sirider/s1/app-development/zhouyi_npu.md b/i18n/en/docusaurus-plugin-content-docs/current/sirider/s1/app-development/zhouyi_npu.md
new file mode 100644
index 000000000..4017c2557
--- /dev/null
+++ b/i18n/en/docusaurus-plugin-content-docs/current/sirider/s1/app-development/zhouyi_npu.md
@@ -0,0 +1,159 @@
+---
+sidebar_position: 1
+---
+
+# Zhouyi Z2 AIPU
+
+The "Zhouyi" AIPU is an innovative AI-specific processor independently developed by Arm China for deep learning. It features an innovative architecture design, providing a complete hardware and software ecosystem with the best balance of PPA (Performance, Power, Area).
+Arm China also provides a range of tools for "Zhouyi" AIPU customers to assist in development, including simulators, compilers, and debuggers for data collection and analysis.
+The "Zhouyi" AIPU supports mainstream AI frameworks, including TensorFlow and ONNX, and will support more extended frameworks in the future.
+
+The "Zhouyi" Z2 AIPU is primarily targeted at high-end security, intelligent cockpits and ADAS (Advanced Driver Assistance Systems), edge servers, and other application scenarios.
+
+## Zhouyi Z2 AIPU User Guide
+
+### Install Zhouyi AIPU SDK on x86 PC
+
+The Zhouyi SDK is a full-stack platform that provides users with rapid development and deployment capabilities.
+
+![image](https://user-images.githubusercontent.com/85479712/198521602-49e13a31-bb49-424f-b782-5108274d63c3.png)
+
+- Download the Zhouyi Z2 SDK installation package from the [Radxa Download Station](https://dl.radxa.com/sirider/s1/) and extract it for installation:
+ ```bash
+ tar -xvf Zhouyi_Z2.tar.gz
+ cd Zhouyi_Z2 && bash +x SETUP.SH
+ ```
+- After installation, the complete SDK files are as follows:
+
+ - `AI610-SDK-r1p3-AIoT`: ARM Zhouyi Z2 toolkit
+
+ - `siengine`: Demos provided by siengine for ARM Zhouyi Z2 model compilation (nn-compiler-user-case-example) and board deployment (nn-runtime-user-case-example)
+
+- Configure the nn-compiler environment:
+ ```bash
+ cd AI610-SDK-r1p3-AIoT/AI610-SDK-r1p3-00eac0/Out-Of-Box/out-of-box-nn-compiler
+ pip3 install -r lib_dependency.txt
+ ```
+ Since this SDK does not include simulation functionality, errors may occur when installing AIPUSimProfiler. These can be ignored.
+
+ If using a virtual environment (venv), please remove the --user option from the pip3 install part in env_setup.sh:
+ ```bash
+ source env_setup.sh
+ ```
+
+### Model Conversion on x86 PC
+
+The nn-compiler can convert models from frameworks like TensorFlow and ONNX into models that can be accelerated by the Zhouyi AIPU for inference.
+
+:::tip
+This case introduces an out-of-the-box example: resnet50 for object classification.
+
+For the complete SDK documentation, please refer to `AI610-SDK-r1p3-AIoT/AI610-SDK-r1p3-00eac0/AI610-DOC-1001-r1p3-eac0`.
+:::
+
+- Enter the siengine nn-compiler-user-case-example directory.
+
+ If the nn-compiler environment is not configured, please follow [Install Zhouyi AIPU SDK on x86 PC](#install-zhouyi-aipu-sdk-on-x86-pc) to configure.
+
+ ```bash
+ cd siengine/nn-compiler-user-case-example/onnx
+ ```
+
+- Generate the quantization calibration set:
+ ```bash
+ python3 generate_calibration_data.py
+ ```
+- Generate image files for model inference:
+ ```bash
+ python3 generate_input_binary.py
+ ```
+ The file is located in ./resnet50/input_3_224_224.bin.
+
+- (Optional) Configure build.cfg (provided in out-of-the-box example):
+ ```bash
+ vim ./resnet50/build.cfg
+ ```
+- Generate the aipu model:
+ ```bash
+ cd ./restnet50
+ aipubuild build.cfg
+ ```
+ The aipu model is generated in ./restnet50 as aipu_mlperf_resnet50.bin.
+
+### Use Zhouyi Z2 for AIPU Model Inference on the Board
+
+Before using Zhouyi Z2 AIPU for inference, a cross-compiled executable file `aiputest` needs to be generated on the x86 host and then copied to the Sirider S1 for execution.
+
+#### Cross-compiling the Binary Executable on x86 PC
+
+- Install the [gcc-linaro-7.5.0-2019.12-x86_64_aarch64-linux-gnu](https://releases.linaro.org/components/toolchain/binaries/latest-7/aarch64-linux-gnu/) cross-compilation toolchain:
+ ```bash
+ tar -xvf gcc-linaro-7.5.0-2019.12-x86_64_aarch64-linux-gnu.tar
+ cp -r gcc-linaro-7.5.0-2019.12-x86_64_aarch64-linux-gnu /opt
+ ```
+- Compile aiputest:
+
+ - Modify the UMDSRC variable:
+ ```bash
+ cd siengine/nn-runtime-user-case-example
+ vim CMakeLists.txt
+ #set(UMDSRC "${CMAKE_SOURCE_DIR}/../AI610-SDK-${AIPU_VERSION}-00eac0/AI610-SDK-1012-${AIPU_VERSION}-eac0/Linux-driver/driver/umd")
+ set(UMDSRC "${CMAKE_SOURCE_DIR}/../../AI610-SDK-${AIPU_VERSION}-AIoT/AI610-SDK-r1p3-00eac0/AI610-SDK-1012-${AIPU_VERSION}-eac0/Linux-driver/driver/umd")
+ ```
+ - Cross-compile:
+ ```bash
+ mkdir build && cd build
+ cmake -DCMAKE_BUILD_TYPE=Release ..
+ make
+ ```
+ The compiled file is located in `siengine/nn-runtime-user-case-example/out/linux/aipu_test`.
+
+#### Inference on the Sirider S1
+
+- Copy the generated `aipu_mlperf_resnet50.bin` model file, `input_3_224_224.bin` image file, `aipu_test` executable file, and `out/linux/libs` dynamic library folder to the Sirider S1.
+- Execute aipu_test:
+ ```bash
+ export LD_LIBRARY_PATH=$LD_LIBRARY_PATH:
+ ./aipu_test aipu_mlperf_resnet50.bin input_3_224_224.bin
+ ```
+ ```bash
+ (aiot-focal_overlayfs)root@linux:~/ssd# ./aipu_test aipu_mlperf_resnet50.bin input_3_224_224.bin
+ usage: ./aipu_test aipu.bin input0.bin
+ aipu_init_context success
+ aipu_load_graph_helper success: aipu_mlperf_resnet50.bin
+ aipu_create_job success
+ Frame #0
+ aipu_finish_job success
+ No profiler data
+ get output tensor 0 success (1/1)
+ output_desc zero_point: 0.0000 scale: 5.5835
+ idx: 637 fval: 21.4919
+ idx: 749 fval: 19.8800
+ idx: 415 fval: 16.1189
+ idx: 412 fval: 15.0443
+ idx: 791 fval: 14.1488
+ Frame #1
+ aipu_finish_job success
+ No profiler data
+ get output tensor 0 success (1/1)
+ output_desc zero_point: 0.0000 scale: 5.5835
+ idx: 637 fval: 21.4919
+ idx: 749 fval: 19.8800
+ idx: 415 fval: 16.1189
+ idx: 412 fval: 15.0443
+ idx: 791 fval: 14.1488
+ aipu_clean_job success
+ aipu_unload_graph success
+ aipu_deinit_ctx success
+ ```
+ The total time for two inferences:
+ ```bash
+ real 0m0.043s
+ user 0m0.008s
+ sys 0m0.023s
+ ```
+
+ The result here only shows the labels of the inference results, with the highest confidence being 637, corresponding to `mailbag, postbag` in [imagenet1000](https://gist.github.com/yrevar/942d3a0ac09ec9e5eb3a).
+
+![input.webp](/img/sirider/s1/aipu_1.webp)
+
diff --git a/static/img/sirider/s1/aipu_1.webp b/static/img/sirider/s1/aipu_1.webp
new file mode 100644
index 000000000..29a0875d7
Binary files /dev/null and b/static/img/sirider/s1/aipu_1.webp differ