Target of the V1.1.0
This version is the first time, we formal release the V1.1.0 of SNNF(Sunplus Neural Network Framework), the original version is V1.
On the basis of v1.0.0.
There are some release notes for the version which are as below:
There are some NN modules and Samples.
lightFace;
age;
lightFace + age;
det_10g;
det_10g + w600k_r50;
humanAttr;
yolov5s + humanAttr;
vehicleAttr;
yolov8s_detection + vehicleAttr;
yolov8s_detection;
yolov8s_obb;
yolov8s_pose;
yolov8s_segment;
yolov8s_classify;
yolov10s_detection;
yolov8n_obb_opti;
yolov8n_pose_opti;
yolov8n_segment_opti;
yolov8n_detection_opti;
yolov8n_classify;
RTMDet-s;
Human tracking;
Human falling detection;
yolov5s_detection;
OCRDet;
OCRCls;
OCRRec;
OCRDet + OCRRec;
OCRDet + OCRCls + OCRRec;
yolov8n_CCPD;
yolov8n_CCPD + OCRRec;
yolov8n_CCPD + OCRCls + OCRRec.
A command mechanism for model list, you can get the all the available models,combination of models and applications and so on, which showed above.
Provide a sample demo for users to refer to, and they can directly run the sample app to verify the NN environment of C3V.
Parallel operations in multi-model applications can maximize the utilization of NPU and CPU, ultimately maximizing the processing frame rate of the application.
Implement official and customized partitions, you can copy an offical model to customized zone for further customizing, or just put your own model in the customized zone as a new model.
Define some specified paths for some resource, for example, config, font, test image, model file, test video and so on.
Define some specified folders such as botSortTrack, imageWriter, videoWriter for assist.
Provide Libs and header files for users to integrate specific NN modules into their applications using the standard interfaces we provide.
Provide COMPILE SH to enable users to easily compile SNNF and locate generated resources in the specified release directory.
Open source the SNNF to facilitate users to better understand the internal operations of SNNF.
Optimize the system to make it more robust, and after long-term testing, the system is stable.
Resource
Please get the V1.0.0 release resource here.
Before starting any work, please carefully read the instruction files such as readme.md in the document directory.
Usage of V1.0.0
How to verify Official Demos
You can use the script we provide to start as follows:
...
, v1.1.0 mainly has the following updates:
Add some NN module.
genderage: gender and age
Det_10g + genderage
LightFace + genderage
Based on the current main framework of SNNF, provide a porting tool so that users can easily add a new model to the Customized zone using the official generated nbg code as the source. The initial version of this tool is currently available, and there will be further updates in the next version.
Fix some known issues, such as:
OCR have a probability of crashing.
HumanAttr have a probability of crashing.
When multiple models are connected in series, if the last model comes out, the result may have an unfinished flag fin=false.
Standardize post-process code.
Organize samples and unittest codes:
samples
is a sample code for snnf command sample which is just for snnf_run.sh.unittest/modelTest
is just for a buildin model's test sample.unittest/pluginTest
is just for the plugin model's test sample.unittest/sequntialTest
is the sample for the pipeline in which several models are sequentially linked together in series.Add NN filter, which can facilitate users to do their own private data processing at the application layer. Users only need to insert the filter directly into the corresponding position of the pipeline when creating the model flow, and then perform private and correct processing at the most appropriate location. The advantage of this is that there is no need to write a dedicated module within SNNF for this private processing, only the official interface of SNNF needs to be called. Of course, SNNF provides some commonly used private processing filters internally, such as label filtering. For example, when Object Detect+HumanAttr, the label results of Object Detect can be filtered, and only the person can be passed to HumanAttr. If the requirements of the application layer happen to match this usage, there is no need to implement it yourself at the application layer. Simply call the methods provided by SNNF.
Provide users with a mAP test sample to compare the precision of the C3V platform's model running results with the precision of the model's official PC running results. The default test model is yolov8s object detection. Users can refer to this mAP sample to rewrite the model they need to test mAP values for.
Improve the toString() function for each model's result, and support the detection result toString JSON format.
Update some make config.
In addition to cross compilation on the server, we also provide compilation methods on the C3V platform.
If it is cross compilation on the server, please confirm the config
CROSS_COMPILE := ON
in makefile_config.mk .If compilation is required on the C3V platform, please confirm the config
CROSS_COMPILE := OFF
in makefile_config.mk .Users can choose the model plugin in makefile_config.mk that needs to be compiled according to their own needs, which can effectively reduce the code size of specific projects and shorten compilation time.
Code Block | ||
---|---|---|
| ||
# ------------------------------------------------------------------------
# define which PLUGIN to build
NN_PLUGIN_ENABLE_YOLOV5S_V2 := yes
NN_PLUGIN_ENABLE_RTMDETS := no
NN_PLUGIN_ENABLE_GEDERAGE := no
NN_PLUGIN_ENABLE_YOLOV8S_DETECTION := no
NN_PLUGIN_ENABLE_YOLOV8S_OBB := no
NN_PLUGIN_ENABLE_YOLOV8S_POSE := no
NN_PLUGIN_ENABLE_YOLOV8S_SEGMENT := no
NN_PLUGIN_ENABLE_YOLOV10S_DETECTION := no
NN_PLUGIN_ENABLE_YOLOV8N_CCPD := no
NN_PLUGIN_ENABLE_YOLOV8N_OBB_OPTI := no
NN_PLUGIN_ENABLE_YOLOV8N_POSE_OPTI := no
NN_PLUGIN_ENABLE_YOLOV8N_SEGMENT_OPTI := no
NN_PLUGIN_ENABLE_YOLOV8N_DETECTION_OPTI := no
NN_PLUGIN_ENABLE_YOLOV8S_DETECTION_OPTI := no
NN_PLUGIN_ENABLE_BOTSORT_TRACK := no |
Adjust the access of certain model header files and provide more necessary header files for the application layer to use.
Change the bbox value type from uint to float. Although this change does not have a substantial impact on the rendering of the results, it can make the precision value more accurate when calculating mAP.
In previous versions, assist modules such as imageWriter/videoWriter/JsonWriter only supported the default filename. Starting from this version, we will support filename config.
Also provide release folder so that new users can easily preview or evaluate our SNNF demo without spending too much time on environment setup and code compilation.
Resource
Please get the V1.1.0 release resource here.
Before starting any work, please carefully read the instruction files such as readme.md in the document directory.
Usage of V1.1.0
How to verify Official Demos
You can use the script we provide to start as follows:
Code Block |
---|
/SNNF/release # ./snnf_run.sh Usage: ./bin/snnf_nnsample [-m|-s|-a|-h] [-i|-v|-o|option] Version: 1.1.0_ Time: [-m,--model <model>] run a single model <model>:Age Det10g HumanAttr HumanFilter LightFace OcrCls OcrDet OcrRec VehicleAttr VehicleFilter W600kR50 YoloV8sOdMap Version: 1.0.0_ Time: Yolov5sDetection [-m,--model <model>] run a single model Yolov5sV1 <model>:Age Yolov8nClassify Det10g Yolov8sClassify stgcn HumanAttr Yolov5sV2 LightFace BotSortTrackStgcn OcrCls GenderAge OcrDet Rtmdets OcrRec YoloV10sDetection YoloV8nCcpdOpti VehicleAttr YoloV8nDetectionBaseOpti W600kR50 YoloV8nDetectionOpti YoloV8nObbOpti Yolov5sDetection YoloV8nPoseOpti Yolov5sV1 Yolov5sV2 YoloV8nSegmentOpti YoloV8sDetection Yolov8nClassify YoloV8sDetectionBaseOpti Yolov8sClassify Yolov5sV3 YoloV8sDetectionOpti YoloV8sObb BotSortTrackStgcn YoloV8sPose Rtmdets YoloV8sSegment YoloV10sDetection example:./bin/snnf_nnsample -m Yolov5sDetection YoloV8nCcpd YoloV8nDetectionBaseOpti ./bin/snnf_nnsample --model HumanAttr YoloV8nDetectionOpti [-s,--sequential <model1,model2,...>] run sequential models YoloV8nObbOpti <models>:Yolov5sDetection,HumanFilter,HumanAttr YoloV8nPoseOpti YoloV8nSegmentOpti LightFace,Age YoloV8sDetection OcrDet,OcrRec YoloV8sDetectionBaseOpti YoloV8sDetectionOpti OcrDet,OcrCls,OcrRec YoloV8sObb YoloV8nCcpdOpti,OcrRec YoloV8sPose YoloV8sSegment Det10g,W600kR50 example:./bin/snnf_nnsample -m Yolov5sDetection YoloV8sDetection,VehicleFilter,VehicleAttr ./bin/snnf_nnsample --model HumanAttr [-s,--sequential <model1,model2,...>] run sequential modelsYoloV8nDetectionOpti,BotSortTrack <models>:Yolov5sDetectionYoloV8nPoseOpti,HumanAttrBotSortTrackStgcn LightFaceYoloV8nCcpdOpti,AgeOcrRec OcrDetYoloV8nCcpdOpti,OcrCls,OcrRec example:./bin/snnf_nnsample OcrDet,OcrCls,OcrRec-s Yolov5sDetection,HumanFilter,HumanAttr ./bin/snnf_nnsample YoloV8nCcpd,OcrRec--sequential ocrDet,ocrCls,ocrRec Det10g,W600kR50 ./bin/snnf_nnsample -s YoloV8nCcpdOpti,OcrRec,imageWriter YoloV8sDetection,VehicleAttr ./bin/snnf_nnsample -s YoloV8sDetectionOpti,BotSortTrack,videoWriter -v resource/video/humanCount.mp4 YoloV8nDetectionOpti,BotSortTrack ./bin/snnf_nnsample -s YoloV8nPoseOpti,BotSortTrackStgcn,videoWriter -v resource/video/person-falling.mp4 [-i,--image file] set image YoloV8nPoseOpti,BotSortTrackStgcnfile to nn detection. <file>: file name YoloV8nCcpd,OcrRec [-c | option]: test count, this parameter is only match YoloV8nCcpd,OcrCls,OcrRecwith -i example:./bin/snnf_nnsample_nnsample -s Yolov5sDetection,HumanAttr -i filename -sc Yolov5s,HumanAttrtestCount ./bin/snnf_nnsample -s Yolov5sDetection,HumanAttr --sequential ocrDet,ocrCls,ocrRecimage filename -c testCount [-v,--video file] set video file to nn detection. ./bin/snnf_nnsample -s YoloV8nCcpd,OcrRec,imageWriter <file>: file name example:./bin/snnf_nnsample -s YoloV8nDetectionOptiYolov5sDetection,BotSortTrack,videoWriterHumanAttr -v resource/video/humanCount.mp4 filename ./bin/snnf_nnsample -s YoloV8nPoseOptiYolov5sDetection,BotSortTrackStgcn,videoWriterHumanAttr -v resource/video/person-falling.mp4-video filename [-io,--imageoutput file] setspecify image file to nn detection. <file>: the output file name for saving results. [-c | option]: test count, this parameter is<file>: onlyfile matchname with -i extension (e.g., output.jpg, output.json, output.mp4) example:./bin/snnf_nnsample -s Yolov5sDetection,HumanAttr -iThis filenameparameter -cmust testCountbe used in conjunction with imageWriter, jsonWriter, or videoWriter. example:./bin/snnf_nnsample -s Yolov5sDetection,HumanAttr,imageWriter --imagei filename -c testCounto output.jpg [-v,--video file] set video file to nn detection. ./bin/snnf_nnsample -s Yolov5sDetection,HumanAttr,jsonWriter -i filename -o output.json <file>: file name example:./bin/snnf_nnsample -s Yolov5sDetection,HumanAttr,videoWriter -v filename -o output.mp4 ./bin/snnf_nnsample -s Yolov5sDetection,HumanAttr --video filename [-a,--all] run all model testing assist tools: imageWriter videoWriter jsonWriter BotSortTrack HumanFilter VehicleFilter |
Release folder structure
...
bin: nnf_nnsample. Prebuild sample programs that can run on the c3v Linux platform.
include:header file of NN framework SDK.
lib:libraries of NN framework SDK.
resource
config: some config files for features.
font: ttf file for plotting sample.
image: image files used for test.
model: models to be used in the sample program.
video: video files used for test.
samples:example code for using NN framework.
snnf_run.sh:executable script for running sample code.
thirdparty: just as its name implies.
...