diff --git a/multipleInferences.md b/multipleInferences.md index 1f16dc5..d295f59 100644 --- a/multipleInferences.md +++ b/multipleInferences.md @@ -4,10 +4,10 @@ How to use multiples GIE's on DeepStream ## 1. Download [my native folder](https://github.com/marcoslucianops/DeepStream-Yolo/tree/master/native), rename to yolo and move to your deepstream/sources folder. -2. Copy each obj.names to deepstream/sources/yolo directory, renaming file to labels_*.txt (* = pgie/sgie1/sgie2/etc), according to each inference type. -3. Copy each yolo.cfg and yolo.weights files to deepstream/sources/yolo directory, renaming files to yolo_*.cfg and yolo_*.weights (* = pgie/sgie1/sgie2/etc), according to each inference type. -4. Make a copy of config_infer_primary.txt file and rename it to config_infer_secondary*.txt (* = 1/2/3/etc), according to inference order. -5. Edit DeepStream for your custom model, according to each yolo_*.cfg (* = pgie/sgie1/sgie2/etc) file: https://github.com/marcoslucianops/DeepStream-Yolo/blob/master/customModels.md +2. Copy each obj.names to deepstream/sources/yolo directory, renaming file to labels_* .txt (* = pgie/sgie1/sgie2/etc), according to each inference type. +3. Copy each yolo.cfg and yolo.weights files to deepstream/sources/yolo directory, renaming files to yolo_* .cfg and yolo_* .weights (* = pgie/sgie1/sgie2/etc), according to each inference type. +4. Make a copy of config_infer_primary.txt file and rename it to config_infer_secondary* .txt (* = 1/2/3/etc), according to inference order. +5. Edit DeepStream for your custom model, according to each yolo_* .cfg (* = pgie/sgie1/sgie2/etc) file: https://github.com/marcoslucianops/DeepStream-Yolo/blob/master/customModels.md **In example folder, on this repository, have all example files to multiple YOLO inferences.** @@ -34,7 +34,7 @@ enable=1 gpu-id=0 gie-unique-id=2 operate-on-gie-id=1 -# If you want secodary inference operate on specified class ids of GIE (class ids you want to operate: 1, 1;2, 2;3;4, 3 etc; comment it if you don't want to use) +# If you want secodary inference operate on specified class ids of GIE (class ids you want to operate: 1, 1;2, 2;3;4, etc; comment it if you don't want to use) operate-on-class-ids=0 nvbuf-memory-type=0 config-file=config_infer_secondary1.txt @@ -144,4 +144,4 @@ To run your custom YOLO model, use this command deepstream-app -c deepstream_app_config.txt ``` -** During test process, engine file will be generated. When engine build process is done, rename engine file according to each configured engine name pgie/sgie1/sgie2/etc) in config_infer file. +**During test process, engine file will be generated. When engine build process is done, rename engine file according to each configured engine name pgie/sgie1/sgie2/etc) in config_infer file.**