GstInference/Benchmarks: Difference between revisions

From RidgeRun Developer Wiki
No edit summary
Tag: Manual revert
 
(173 intermediate revisions by 6 users not shown)
Line 1: Line 1:
<noinclude>
<noinclude>
{{GstInference/Head|previous=Example Applications/DispTec|next=Model Zoo|keywords=GstInference gstreamer pipelines, Inference gstreamer pipelines, NCSDK Inference gstreamer pipelines, GoogLeNet, TinyYolo v2, GoogLeNet x2, TensorFlow backend|title=GstInference Benchmarks}}
{{GstInference/Head|previous=Example Applications/DispTec|next=Model Zoo|metakeywords=GstInference gstreamer pipelines, Inference gstreamer pipelines, NCSDK Inference gstreamer pipelines, GoogLeNet, TinyYolo v2, GoogLeNet x2, TensorFlow backend|title=GstInference Benchmarks}}
</noinclude>
</noinclude>


== GstInference Benchmarks ==
 
The following benchmarks were run with a source video (1920x1080@60). With the following base [https://www.ridgerun.com/gstreamer GStreamer] pipeline, and environment variables:
<html>
<script
/* This should be added once per page only */
type="text/javascript" src="https://www.gstatic.com/charts/loader.js">
</script>
</html>
 
== GstInference Benchmarks - Introduction ==
 
This wiki summarizes a series of benchmarks on different hardware platforms based on the [https://github.com/RidgeRun/gst-inference/blob/master/tests/benchmark/run_benchmark.sh run_benchmark.sh] bash script that can be found in the official [https://github.com/RidgeRun/gst-inference GstInference repository]. The script is based on the following GStreamer pipeline:


<source lang="bash">
<source lang="bash">
$ VIDEO_FILE='video.mp4'
#Script to run each model
$ MODEL_LOCATION='graph_inceptionv1_tensorflow.pb'
run_all_models(){
$ INPUT_LAYER='input'
 
$ OUTPUT_LAYER='InceptionV1/Logits/Predictions/Reshape_1'
  model_array=(inceptionv1 inceptionv2 inceptionv3 inceptionv4 tinyyolov2 tinyyolov3)
</source>
  model_upper_array=(InceptionV1 InceptionV2 InceptionV3 InceptionV4 TinyYoloV2 TinyYoloV3)
The environment variables were changed accordingly with the used model (Inception V1,V2,V3 or V4)
  input_array=(input input input input input/Placeholder inputs )
  output_array=(InceptionV1/Logits/Predictions/Reshape_1 Softmax InceptionV3/Predictions/Reshape_1
  InceptionV4/Logits/Predictions add_8 output_boxes )
 
  mkdir -p logs/
  rm -f logs/*
 
  for ((i=0;i<${#model_array[@]};++i)); do
    echo Perf ${model_array[i]}
    gst-launch-1.0 \
    filesrc location=$VIDEO_PATH num-buffers=600 ! decodebin ! videoconvert ! \
    perf print-arm-load=true name=inputperf ! tee name=t t. ! videoscale ! queue ! net.sink_model t. ! queue ! net.sink_bypass \
    ${model_array[i]} backend=$BACKEND name=net backend::input-layer=${input_array[i]} backend::output-layer=${output_array[i]} \
    model-location="${MODELS_PATH}${model_upper_array[i]}_${INTERNAL_PATH}/graph_${model_array[i]}${EXTENSION}" \
    net.src_bypass ! perf print-arm-load=true name=outputperf ! videoconvert ! fakesink sync=false > logs/${model_array[i]}.log
  done
}
</source>  
 
=== Test benchmark video ===
The following video was used to perform the benchmark tests.
<br>
To download the video press right-click on the video and select 'Save video as' and save this on your computer.
<br>
<br>
[[File:Test benchmark video.mp4|thumb|border|center|500px|alt=Alt|Video 1. Test benchmark video]]


<source lang="bash">
== x86 ==
GST_DEBUG=inception1:1 gst-launch-1.0 filesrc location=$VIDEO_FILE ! decodebin ! videoconvert ! videoscale ! queue ! net.sink_model inceptionv1 name=net model-location=$MODEL_LOCATION backend=tensorflow backend::input-layer=$INPUT_LAYER  backend::output-layer=$OUTPUT_LAYER net.src_model ! perf ! fakesink -v
</source>


The Desktop PC had the following specifications:
The Desktop PC had the following specifications:
*Intel(R) Core(TM) i7-3770 CPU @ 3.40GHz
*Intel(R) Core(TM) Core i7-7700HQ CPU @ 2.80GHz
*8 GB RAM
*12 GB RAM
*Cedar [Radeon HD 5000/6000/7350/8350 Series]
*Linux 4.15.0-106-generic x86_64 (Ubuntu 16.04)
*Linux 4.15.0-54-generic x86_64 (Ubuntu 16.04)
*GStreamer 1.8.3
 
=== FPS Measurements ===
 
<html>
 
<style>
    .button {
    background-color: #008CBA;
    border: none;
    color: white;
    padding: 15px 32px;
    text-align: center;
    text-decoration: none;
    display: inline-block;
    font-size: 16px;
    margin: 4px 2px;
    cursor: pointer;
  }
</style>
 
<div id="chart_fps_x86" style="margin: auto; width: 800px; height: 500px;"></div>
 
<script>
      google.charts.load('current', {'packages':['corechart', 'bar']});
      google.charts.setOnLoadCallback(drawStuffx86Fps);
     
      function drawStuffx86Fps() {


The Jetson Xavier power modes used were 2 and 6 (more information: [https://docs.nvidia.com/jetson/l4t/index.html#page/Tegra%2520Linux%2520Driver%2520Package%2520Development%2520Guide%2Fpower_management_jetson_xavier.html%23wwpID0E0OM0HA Supported Modes and Power Efficiency] )
        var chartDiv_Fps_x86 = document.getElementById('chart_fps_x86');
 
        var table_models_fps_x86 = google.visualization.arrayToDataTable([
          ['Model',                      //Column 0
          'ONNXRT \n x86',
          'ONNXRT OpenVINO (CPU_FP32) \n x86',
          'ONNXRT OpenVINO (GPU_FP32)\n x86',
          'ONNXRT OpenVINO (GPU_FP16)\n x86',
          'ONNXRT OpenVINO (MYRIAD_FP16)\n x86',
          'TensorFlow \n x86',
          'TensorFlow Lite \n x86'],
          ['InceptionV1', 47.9, 81.966, 70.580, 98.742, 46.294, 55.3182, 18.8422], //row 1
          ['InceptionV2', 32.7, 63.352, 54.159, 77.449, 34.613, 39.6438, 13.5714], //row 2
          ['InceptionV3', 12.1, 23.287, 20.878, 34.059, 11.999, 16.2488, 4.9924], //row 3
          ['InceptionV4', 5.26, 10.927, 6.160, 4.548, 6.494, 7.793, 2.583], //row 4
          ['TinyYoloV2', 33.559, 32.587, 0, 0, 0, 18.1846, 7.2708], //row 5
          ['TinyYoloV3', 35.092, 27.799, 0, 0, 0, 21.7334, 7.3042]  //row 6
        ]);
        var x86_materialOptions_fps = {
          width: 1000,
          chart: {
            title: 'Model Vs FPS per backend',
          },
          series: {
          },
          axes: {
            y: {
              distance: {side: 'left',label: 'FPS'}, // Left y-axis.
            }
          }
        };
 
        var materialChart_x86_fps = new google.charts.Bar(chartDiv_Fps_x86);
        view_x86_fps = new google.visualization.DataView(table_models_fps_x86);
 
        function drawMaterialChart() {
          var materialChart_x86_fps = new google.charts.Bar(chartDiv_Fps_x86);
          materialChart_x86_fps.draw(table_models_fps_x86, google.charts.Bar.convertOptions(x86_materialOptions_fps));
 
          init_charts();
        }
        function init_charts(){
          view_x86_fps.setColumns([0,1, 2, 3, 4, 5, 6, 7]);
          materialChart_x86_fps.draw(view_x86_fps, x86_materialOptions_fps);
        }
        drawMaterialChart();
        }
 
</script>
 
</html>
 
=== CPU Load Measurements ===
 
<html>
 
<style>
    .button {
    background-color: #008CBA;
    border: none;
    color: white;
    padding: 15px 32px;
    text-align: center;
    text-decoration: none;
    display: inline-block;
    font-size: 16px;
    margin: 4px 2px;
    cursor: pointer;
  }
</style>
 
<div id="chart_cpu_x86" style="margin: auto; width: 800px; height: 500px;"></div>
 
<script>
      google.charts.load('current', {'packages':['corechart', 'bar']});
      google.charts.setOnLoadCallback(drawStuffx86Cpu);
     
      function drawStuffx86Cpu() {
 
        var chartDiv_Cpu_x86 = document.getElementById('chart_cpu_x86');
 
        var table_models_cpu_x86 = google.visualization.arrayToDataTable([
          ['Model',                      //Column 0
          'ONNXRT \n x86',
          'ONNXRT OpenVINO (CPU_FP32) \n x86',
          'ONNXRT OpenVINO (GPU_FP32)\n x86',
          'ONNXRT OpenVINO (GPU_FP16)\n x86',
          'ONNXRT OpenVINO (MYRIAD_FP16)\n x86',
          'TensorFlow \n x86',
          'TensorFlow Lite \n x86'],        //Column 1
          ['InceptionV1', 94.6, 49, 31, 29, 14, 74.2, 47.6], //row 1
          ['InceptionV2', 100, 52, 28, 29, 11, 74.2, 43.6], //row 2
          ['InceptionV3', 95.2, 49, 28, 28, 13, 81, 60.2], //row 3
          ['InceptionV4', 88.8, 49, 33, 46, 11, 86, 50], //row 4
          ['TinyYoloV2',  94, 50, 0, 0, 0, 80.6, 46], //row 5
          ['TinyYoloV3',  91.4, 46, 0, 0, 0, 74.6, 42.4]  //row 6
        ]);
        var x86_materialOptions_cpu = {
          width: 1000,
          chart: {
            title: 'Model Vs CPU Load per backend',
          },
          series: {
          },
          axes: {
            y: {
              distance: {side: 'left',label: 'CPU Load'}, // Left y-axis.
            }
          }
        };
 
        var materialChart_x86_cpu = new google.charts.Bar(chartDiv_Cpu_x86);
        view_x86_cpu = new google.visualization.DataView(table_models_cpu_x86);
 
        function drawMaterialChart() {
          var materialChart_x86_cpu = new google.charts.Bar(chartDiv_Cpu_x86);
          materialChart_x86_cpu.draw(table_models_cpu_x86, google.charts.Bar.convertOptions(x86_materialOptions_cpu));
 
          init_charts();
        }
        function init_charts(){
          view_x86_cpu.setColumns([0,1, 2, 3, 4, 5, 6, 7]);
          materialChart_x86_cpu.draw(view_x86_cpu, x86_materialOptions_cpu);
        }
        drawMaterialChart();
        }
 
</script>
 
</html>
 
== Jetson AGX Xavier ==
 
The Jetson Xavier power modes used were 2 and 6 (more information: [https://docs.nvidia.com/jetson/l4t/index.html#page/Tegra%2520Linux%2520Driver%2520Package%2520Development%2520Guide%2Fpower_management_jetson_xavier.html%23wwpID0E0OM0HA Supported Modes and Power Efficiency])


*View current power mode:
*View current power mode:
Line 39: Line 231:
Where x is the power mode ID (e.g. 0, 1, 2, 3, 4, 5, 6).
Where x is the power mode ID (e.g. 0, 1, 2, 3, 4, 5, 6).


=== Summary ===  
=== FPS Measurements ===


{| class="wikitable" style="display: inline-table;"
<html>
! style="font-weight:bold; background-color:#efefef; color:#000000;" | Desktop PC
! colspan="2" style="text-align: center; font-weight:bold; background-color:#efefef; color:#000000;" | CPU Library
|-
| style="background-color:#e98d44; color:#000000;" | Model
| style="background-color:#e98d44; color:#000000;" | Framerate
| style="background-color:#e98d44; color:#000000;" | CPU Usage
|-
| style="background-color:#e98d44; color:#000000;" | Inception V1
| style="background-color:#fee3cd; color:#000000;" | 11.89
| style="background-color:#fee3cd; color:#000000;" | 48
|-
| style="background-color:#e98d44; color:#000000;" | Inception V2
| style="background-color:#fee3cd; color:#000000;" | 10.33
| style="background-color:#fee3cd; color:#000000;" | 65
|-
| style="background-color:#e98d44; color:#000000;" | Inception V3
| style="background-color:#fee3cd; color:#000000;" | 5.41
| style="background-color:#fee3cd; color:#000000;" | 90
|-
| style="background-color:#e98d44; color:#000000;" | Inception V4
| style="background-color:#fee3cd; color:#000000;" | 3.81
| style="background-color:#fee3cd; color:#000000;" | 94
|}


{| class="wikitable" style="display: inline-table;"
<style>
! style="font-weight:bold; background-color:#efefef; color:#000000;" | Jetson Xavier (15W)
    .button {
! colspan="2" style="text-align: center; font-weight:bold; background-color:#efefef; color:#000000;" | CPU Library
    background-color: #008CBA;
! colspan="2" style="text-align: center; font-weight:bold; background-color:#efefef; color:#000000;" | GPU Library
    border: none;
|-
    color: white;
| style="background-color:#2c79d3; color:#000000;" | Model
    padding: 15px 32px;
| style="background-color:#2c79d3; color:#000000;" | Framerate
    text-align: center;
| style="background-color:#2c79d3; color:#000000;" | CPU Usage
    text-decoration: none;
| style="background-color:#2c79d3; color:#000000;" | Framerate
    display: inline-block;
| style="background-color:#2c79d3; color:#000000;" | CPU Usage
    font-size: 16px;
|-
    margin: 4px 2px;
| style="background-color:#2c79d3; color:#000000;" | Inception V1
    cursor: pointer;
| style="background-color:#c5daf6; color:#000000;" | 8.24
  }
| style="background-color:#c5daf6; color:#000000;" | 86
</style>
| style="background-color:#c5daf6; color:#000000;" | 52.3
| style="background-color:#c5daf6; color:#000000;" | 43
|-
| style="background-color:#2c79d3; color:#000000;" | Inception V2
| style="background-color:#c5daf6; color:#000000;" | 6.58
| style="background-color:#c5daf6; color:#000000;" | 88
| style="background-color:#c5daf6; color:#000000;" | 39.6
| style="background-color:#c5daf6; color:#000000;" | 42
|-
| style="background-color:#2c79d3; color:#000000;" | Inception V3
| style="background-color:#c5daf6; color:#000000;" | 2.54
| style="background-color:#c5daf6; color:#000000;" | 92
| style="background-color:#c5daf6; color:#000000;" | 17.8
| style="background-color:#c5daf6; color:#000000;" | 25
|-
| style="background-color:#2c79d3; color:#000000;" | Inception V4
| style="background-color:#c5daf6; color:#000000;" | 1.22
| style="background-color:#c5daf6; color:#000000;" | 94
| style="background-color:#c5daf6; color:#000000;" | 9.4
| style="background-color:#c5daf6; color:#000000;" | 20
|}


{| class="wikitable" style="display: inline-table;"
<div id="chart_fps_xavier" style="margin: auto; width: 800px; height: 500px;"></div>
! style="font-weight:bold; background-color:#efefef; color:#000000;" | Jetson Xavier (30W)
! colspan="2" style="text-align: center; font-weight:bold; background-color:#efefef; color:#000000;" | CPU Library
! colspan="2" style="text-align: center; font-weight:bold; background-color:#efefef; color:#000000;" | GPU Library
|-
| style="background-color:#6aa758; color:#000000;" | Model
| style="background-color:#6aa758; color:#000000;" | Framerate
| style="background-color:#6aa758; color:#000000;" | CPU Usage
| style="background-color:#6aa758; color:#000000;" | Framerate
| style="background-color:#6aa758; color:#000000;" | CPU Usage
|-
| style="background-color:#6aa758; color:#000000;" | Inception V1
| style="background-color:#d8e9d3; color:#000000;" | 6.41
| style="background-color:#d8e9d3; color:#000000;" | 93
| style="background-color:#d8e9d3; color:#000000;" | 66.27
| style="background-color:#d8e9d3; color:#000000;" | 72
|-
| style="background-color:#6aa758; color:#000000;" | Inception V2
| style="background-color:#d8e9d3; color:#000000;" | 5.11
| style="background-color:#d8e9d3; color:#000000;" | 95
| style="background-color:#d8e9d3; color:#000000;" | 50.59
| style="background-color:#d8e9d3; color:#000000;" | 62
|-
| style="background-color:#6aa758; color:#000000;" | Inception V3
| style="background-color:#d8e9d3; color:#000000;" | 1.96
| style="background-color:#d8e9d3; color:#000000;" | 98
| style="background-color:#d8e9d3; color:#000000;" | 22.95
| style="background-color:#d8e9d3; color:#000000;" | 44
|-
| style="background-color:#6aa758; color:#000000;" | Inception V4
| style="background-color:#d8e9d3; color:#000000;" | 0.98
| style="background-color:#d8e9d3; color:#000000;" | 99
| style="background-color:#d8e9d3; color:#000000;" | 12.14
| style="background-color:#d8e9d3; color:#000000;" | 32
|}


=== Framerate ===
<script>
      google.charts.load('current', {'packages':['corechart', 'bar']});
      google.charts.setOnLoadCallback(drawStuffXavierFps);
     
      function drawStuffXavierFps() {


[[File:Framerate Benchmarks gst-inference.png|1024px|frameless|thumb|center]]
        var chartDiv_Fps_Xavier = document.getElementById('chart_fps_xavier');


=== CPU Usage ===
        var table_models_fps_xavier = google.visualization.arrayToDataTable([
          ['Model',                      //Column 0
          'TensorFlow \n Xavier (15 W)',      //Column 1
          'TensorFlow (GPU) \n Xavier (15 W)', //Column 2
          'TensorFlow \n Xavier (30 W)',      //Column 3
          'TensorFlow (GPU) \n Xavier (30 W)', //Column 4
          'TensorRT \n Xavier',              //Column 5
          'ONNXRT ACL \n Xavier'],            //Column 6
          ['InceptionV1', 8.24, 52.3, 6.41, 66.27, 92.6, 17.566], //row 1
          ['InceptionV2', 6.58, 39.6, 5.11, 50.59, 0, 12.729], //row 2
          ['InceptionV3', 2.54, 17.8, 1.96, 22.95, 24.9, 5.709], //row 3
          ['InceptionV4', 1.22, 9.4, 0.98, 12.14, 13.6, 2.747], //row 4
          ['TinyYoloV2',  0, 0, 0, 0, 69.7, 9.367], //row 5
          ['TinyYoloV3',  0, 0, 0, 0, 0, 10.520]  //row 6
        ]);
        var xavier_materialOptions_fps = {
          width: 900,
          chart: {
            title: 'Model Vs FPS per backend',
          },
          series: {
          },
          axes: {
            y: {
              distance: {side: 'left',label: 'FPS'}, // Left y-axis.
            }
          }
        };


[[File:CPU Benchmarks gst-inference.png|1024px|frameless|thumb|center]]
        var materialChart_xavier_fps = new google.charts.Bar(chartDiv_Fps_Xavier);
        view_xavier_fps = new google.visualization.DataView(table_models_fps_xavier);
 
        function drawMaterialChart() {
          var materialChart_xavier_fps = new google.charts.Bar(chartDiv_Fps_Xavier);
          materialChart_xavier_fps.draw(table_models_fps_xavier, google.charts.Bar.convertOptions(xavier_materialOptions_fps));
 
          init_charts();
        }
        function init_charts(){
          view_xavier_fps.setColumns([0,1,2,3,4,5,6]);
          materialChart_xavier_fps.draw(view_xavier_fps, xavier_materialOptions_fps);
        }
        drawMaterialChart();
        }
 
</script>
 
</html>
 
=== CPU Load Measurements ===


== TensorFlow Lite Benchmarks ==
<html>
<html>
  <script type="text/javascript" src="https://www.gstatic.com/charts/loader.js"></script>
 
    
<style>
<br><br>
    .button {
<div id="chart_div" style="width: 800px; height: 500px;"></div>
    background-color: #008CBA;
    border: none;
    color: white;
    padding: 15px 32px;
    text-align: center;
    text-decoration: none;
    display: inline-block;
    font-size: 16px;
    margin: 4px 2px;
    cursor: pointer;
   }
</style>
 
<div id="chart_cpu_xavier" style="margin: auto; width: 800px; height: 500px;"></div>


<script>
<script>
        google.charts.load('current', {'packages':['corechart', 'bar']});
      google.charts.load('current', {'packages':['corechart', 'bar']});
       google.charts.setOnLoadCallback(drawStuff);
       google.charts.setOnLoadCallback(drawStuffXavierCpu);
     
      function drawStuffXavierCpu() {
 
        var chartDiv_Cpu_Xavier = document.getElementById('chart_cpu_xavier');
 
        var table_models_cpu_xavier = google.visualization.arrayToDataTable([
          ['Model',                      //Column 0
          'TensorFlow \n Xavier (15 W)',      //Column 1
          'TensorFlow (GPU) \n Xavier (15 W)', //Column 2
          'TensorFlow \n Xavier (30 W)',      //Column 3
          'TensorFlow (GPU) \n Xavier (30 W)', //Column 4
          'TensorRT \n Xavier',              //Column 5
          'ONNXRT ACL \n Xavier'],            //Column 6
          ['InceptionV1', 86, 72, 93, 72, 32, 50], //row 1
          ['InceptionV2', 88, 62.6, 95, 62, 0, 49], //row 2
          ['InceptionV3', 92, 44, 98, 44, 6, 50], //row 3
          ['InceptionV4', 94, 32, 99, 32, 3, 50], //row 4
          ['TinyYoloV2',  0, 0, 0, 0, 16, 50], //row 5
          ['TinyYoloV3',  0, 0, 0, 0, 0, 50]  //row 6
        ]);
        var xavier_materialOptions_cpu = {
          width: 900,
          chart: {
            title: 'Model Vs CPU load per backend',
          },
          series: {
          },
          axes: {
            y: {
              distance: {side: 'left',label: 'CPU Load'}, // Left y-axis.
            }
          }
        };
 
        var materialChart_xavier_cpu = new google.charts.Bar(chartDiv_Cpu_Xavier);
        view_xavier_cpu = new google.visualization.DataView(table_models_cpu_xavier);


      function drawStuff() {
        function drawMaterialChart() {
          var materialChart_xavier_cpu = new google.charts.Bar(chartDiv_Cpu_Xavier);
          materialChart_xavier_cpu.draw(table_models_cpu_xavier, google.charts.Bar.convertOptions(xavier_materialOptions_cpu));


         var button = document.getElementById('change-chart');
          init_charts();
         var chartDiv = document.getElementById('chart_div');
        }
         function init_charts(){
          view_xavier_cpu.setColumns([0,1,2,3,4,5,6]);
          materialChart_xavier_cpu.draw(view_xavier_cpu, xavier_materialOptions_cpu);
         }
        drawMaterialChart();
        }


         var data = google.visualization.arrayToDataTable([
</script>
           ['Model', 'FPS', 'CPU'],
 
           ['InceptionV1', 21, 50],
</html>
           ['InceptionV2', 16, 45],
 
           ['InceptionV3', 6, 32],
== Jetson TX2 ==
           ['InceptionV4', 3, 15],
 
           ['TinyyoloV2', 9, 49],
=== FPS Measurements ===
           ['TinyyoloV2', 10, 58]
 
<html>
 
<style>
    .button {
    background-color: #008CBA;
    border: none;
    color: white;
    padding: 15px 32px;
    text-align: center;
    text-decoration: none;
    display: inline-block;
    font-size: 16px;
    margin: 4px 2px;
    cursor: pointer;
  }
</style>
 
<div id="chart_fps_tx2" style="margin: auto; width: 800px; height: 500px;"></div>
 
<script>
      google.charts.load('current', {'packages':['corechart', 'bar']});
      google.charts.setOnLoadCallback(drawStuffTx2Fps);
     
      function drawStuffTx2Fps() {
 
        var chartDiv_Fps_Tx2 = document.getElementById('chart_fps_tx2');
 
         var table_models_fps_tx2 = google.visualization.arrayToDataTable([
           ['Model',                     //Column 0
          'TensorRT \n TX2'],       //Column 1
           ['InceptionV1', 62.5], //row 1
           ['InceptionV2', 0], //row 2
           ['InceptionV3', 20.2], //row 3
           ['InceptionV4', 11.22], //row 4
           ['TinyYoloV2', 33.8], //row 5
           ['TinyYoloV3', 0] //row 6
         ]);
         ]);
        var Tx2_materialOptions_fps = {
          width: 900,
          chart: {
            title: 'Model Vs FPS per backend',
          },
          series: {
          },
          axes: {
            y: {
              distance: {side: 'left',label: 'FPS'}, // Left y-axis.
            }
          }
        };


         var materialOptions = {
         var materialChart_tx2_fps = new google.charts.Bar(chartDiv_Fps_Tx2);
        view_tx2_fps = new google.visualization.DataView(table_models_fps_tx2);
 
        function drawMaterialChart() {
          var materialChart_tx2_fps = new google.charts.Bar(chartDiv_Fps_Tx2);
          materialChart_tx2_fps.draw(table_models_fps_tx2, google.charts.Bar.convertOptions(Tx2_materialOptions_fps));
 
          init_charts();
        }
        function init_charts(){
          view_tx2_fps.setColumns([0,1]);
          materialChart_tx2_fps.draw(view_tx2_fps, Tx2_materialOptions_fps);
        }
        drawMaterialChart();
        }
 
</script>
 
</html>
 
=== CPU Load Measurements ===
 
<html>
 
<style>
    .button {
    background-color: #008CBA;
    border: none;
    color: white;
    padding: 15px 32px;
    text-align: center;
    text-decoration: none;
    display: inline-block;
    font-size: 16px;
    margin: 4px 2px;
    cursor: pointer;
  }
</style>
 
<div id="chart_cpu_tx2" style="margin: auto; width: 800px; height: 500px;"></div>
 
<script>
      google.charts.load('current', {'packages':['corechart', 'bar']});
      google.charts.setOnLoadCallback(drawStuffTx2Cpu);
     
      function drawStuffTx2Cpu() {
 
        var chartDiv_Cpu_Tx2 = document.getElementById('chart_cpu_tx2');
 
        var table_models_cpu_tx2 = google.visualization.arrayToDataTable([
          ['Model',                      //Column 0
          'TensorRT \n TX2'],        //Column 1
          ['InceptionV1', 21], //row 1
          ['InceptionV2', 0], //row 2
          ['InceptionV3', 10], //row 3
          ['InceptionV4', 7], //row 4
          ['TinyYoloV2',  19], //row 5
          ['TinyYoloV3',  0]  //row 6
        ]);
        var Tx2_materialOptions_cpu = {
          width: 900,
          chart: {
            title: 'Model Vs CPU Load per backend',
          },
          series: {
          },
          axes: {
            y: {
              distance: {side: 'left',label: 'FPS'}, // Left y-axis.
            }
          }
        };
 
        var materialChart_tx2_cpu = new google.charts.Bar(chartDiv_Cpu_Tx2);
        view_tx2_cpu = new google.visualization.DataView(table_models_cpu_tx2);
 
        function drawMaterialChart() {
          var materialChart_tx2_cpu = new google.charts.Bar(chartDiv_Cpu_Tx2);
          materialChart_tx2_cpu.draw(table_models_cpu_tx2, google.charts.Bar.convertOptions(Tx2_materialOptions_cpu));
 
          init_charts();
        }
        function init_charts(){
          view_tx2_cpu.setColumns([0,1]);
          materialChart_tx2_cpu.draw(view_tx2_cpu, Tx2_materialOptions_cpu);
        }
        drawMaterialChart();
        }
 
</script>
 
</html>
 
== Jetson Nano ==
 
=== FPS Measurements ===
 
<html>
 
<style>
    .button {
    background-color: #008CBA;
    border: none;
    color: white;
    padding: 15px 32px;
    text-align: center;
    text-decoration: none;
    display: inline-block;
    font-size: 16px;
    margin: 4px 2px;
    cursor: pointer;
  }
</style>
 
<div id="chart_fps_nano" style="margin: auto; width: 800px; height: 500px;"></div>
 
<script>
      google.charts.load('current', {'packages':['corechart', 'bar']});
      google.charts.setOnLoadCallback(drawStuffNanoFps);
     
      function drawStuffNanoFps() {
 
        var chartDiv_Fps_Nano = document.getElementById('chart_fps_nano');
 
        var table_models_fps_nano = google.visualization.arrayToDataTable([
          ['Model',                      //Column 0
          'TensorRT \n Nano'],        //Column 1
          ['InceptionV1', 36.8], //row 1
          ['InceptionV2', 0], //row 2
          ['InceptionV3', 10.9], //row 3
          ['InceptionV4', 4.82], //row 4
          ['TinyYoloV2',  14.6], //row 5
          ['TinyYoloV3',  0]  //row 6
        ]);
        var Nano_materialOptions_fps = {
           width: 900,
           width: 900,
           chart: {
           chart: {
             title: 'Model performance',
             title: 'Model Vs FPS per backend',
            subtitle: 'distance on the left, brightness on the right'
           },
           },
           series: {
           series: {
            0: { axis: 'FPS' }, // Bind series 0 to an axis named 'distance'.
            1: { axis: 'CPU Usage %' } // Bind series 1 to an axis named 'brightness'.
           },
           },
           axes: {
           axes: {
             y: {
             y: {
               distance: {label: 'FPS'}, // Left y-axis.
               distance: {side: 'left',label: 'FPS'}, // Left y-axis.
              brightness: {side: 'right', label: 'CPU Usage %'} // Right y-axis.
             }
             }
           }
           }
         };
         };


        var materialChart_nano_fps = new google.charts.Bar(chartDiv_Fps_Nano);
        view_nano_fps = new google.visualization.DataView(table_models_fps_nano);
        function drawMaterialChart() {
          var materialChart_nano_fps = new google.charts.Bar(chartDiv_Fps_Nano);
          materialChart_nano_fps.draw(table_models_fps_nano, google.charts.Bar.convertOptions(Nano_materialOptions_fps));
          init_charts();
        }
        function init_charts(){
          view_nano_fps.setColumns([0,1]);
          materialChart_nano_fps.draw(view_nano_fps, Nano_materialOptions_fps);
        }
        drawMaterialChart();
        }
</script>
</html>
=== CPU Load Measurements ===
<html>
<style>
    .button {
    background-color: #008CBA;
    border: none;
    color: white;
    padding: 15px 32px;
    text-align: center;
    text-decoration: none;
    display: inline-block;
    font-size: 16px;
    margin: 4px 2px;
    cursor: pointer;
  }
</style>
<div id="chart_cpu_nano" style="margin: auto; width: 800px; height: 500px;"></div>
<script>
      google.charts.load('current', {'packages':['corechart', 'bar']});
      google.charts.setOnLoadCallback(drawStuffNanoCpu);
     
      function drawStuffNanoCpu() {
        var chartDiv_Cpu_Nano = document.getElementById('chart_cpu_nano');
        var table_models_cpu_nano = google.visualization.arrayToDataTable([
          ['Model',                      //Column 0
          'TensorRT \n Nano'],        //Column 1
          ['InceptionV1', 18], //row 1
          ['InceptionV2', 0], //row 2
          ['InceptionV3', 18], //row 3
          ['InceptionV4', 20], //row 4
          ['TinyYoloV2',  20], //row 5
          ['TinyYoloV3',  0]  //row 6
        ]);
        var Nano_materialOptions_cpu = {
          width: 900,
          chart: {
            title: 'Model Vs CPU Load per backend',
          },
          series: {
          },
          axes: {
            y: {
              distance: {side: 'left',label: 'FPS'}, // Left y-axis.
            }
          }
        };


        var materialChart_nano_cpu = new google.charts.Bar(chartDiv_Cpu_Nano);
        view_nano_cpu = new google.visualization.DataView(table_models_cpu_nano);


         function drawMaterialChart() {
         function drawMaterialChart() {
           var materialChart = new google.charts.Bar(chartDiv);
           var materialChart_nano_cpu = new google.charts.Bar(chartDiv_Cpu_Nano);
           materialChart.draw(data, google.charts.Bar.convertOptions(materialOptions));
           materialChart_nano_cpu.draw(table_models_cpu_nano, google.charts.Bar.convertOptions(Nano_materialOptions_cpu));


          init_charts();
         }
         }
        function init_charts(){
          view_nano_cpu.setColumns([0,1]);
          materialChart_nano_cpu.draw(view_nano_cpu, Nano_materialOptions_cpu);
        }
        drawMaterialChart();
        }
</script>
</html>
== Google Coral ==
The following benchmarks were performed on the Coral Dev Board.
=== FPS Measurements ===
<html>
<style>
    .button {
    background-color: #008CBA;
    border: none;
    color: white;
    padding: 15px 32px;
    text-align: center;
    text-decoration: none;
    display: inline-block;
    font-size: 16px;
    margin: 4px 2px;
    cursor: pointer;
  }
</style>
<div id="chart_fps_coral" style="margin: auto; width: 800px; height: 500px;"></div>
<script>
      google.charts.load('current', {'packages':['corechart', 'bar']});
      google.charts.setOnLoadCallback(drawStuffCoralFps);
     
      function drawStuffCoralFps() {
        var chartDiv_Fps_Coral = document.getElementById('chart_fps_coral');
        var table_models_fps_coral = google.visualization.arrayToDataTable([
          ['Model',                      //Column 0
          'TensorFlow Lite \n Coral',
          'TensorFlow Lite EdgeTPU \n Coral'],        //Column 1
          ['InceptionV1', 3.11, 41.6], //row 1
          ['InceptionV2', 2.31, 42.8], //row 2
          ['InceptionV3', 0.9, 15.02], //row 3
          ['InceptionV4', 0, 8.56], //row 4
          ['MobileNetV2',  0, 41.12], //row 5
          ['MobileNetV2 + SSD',  0, 38.64]  //row 6
        ]);
        var Coral_materialOptions_fps = {
          width: 900,
          chart: {
            title: 'Model Vs FPS per backend',
          },
          series: {
          },
          axes: {
            y: {
              distance: {side: 'left',label: 'FPS'}, // Left y-axis.
            }
          }
        };
        var materialChart_coral_fps = new google.charts.Bar(chartDiv_Fps_Coral);
        view_coral_fps = new google.visualization.DataView(table_models_fps_coral);


         function drawClassicChart() {
         function drawMaterialChart() {
           var classicChart = new google.visualization.ColumnChart(chartDiv);
           var materialChart_coral_fps = new google.charts.Bar(chartDiv_Fps_Coral);
           classicChart.draw(data, classicOptions);
           materialChart_coral_fps.draw(table_models_fps_coral, google.charts.Bar.convertOptions(Coral_materialOptions_fps));


          init_charts();
         }
         }
        function init_charts(){
          view_coral_fps.setColumns([0,1, 2]);
          materialChart_coral_fps.draw(view_coral_fps, Coral_materialOptions_fps);
        }
        drawMaterialChart();
        }
</script>


</html>
=== CPU Load Measurements ===
<html>
<style>
    .button {
    background-color: #008CBA;
    border: none;
    color: white;
    padding: 15px 32px;
    text-align: center;
    text-decoration: none;
    display: inline-block;
    font-size: 16px;
    margin: 4px 2px;
    cursor: pointer;
  }
</style>
<div id="chart_cpu_coral" style="margin: auto; width: 800px; height: 500px;"></div>
<script>
      google.charts.load('current', {'packages':['corechart', 'bar']});
      google.charts.setOnLoadCallback(drawStuffCoralCpu);
     
      function drawStuffCoralCpu() {
        var chartDiv_Cpu_Coral = document.getElementById('chart_cpu_coral');
        var table_models_cpu_coral = google.visualization.arrayToDataTable([
          ['Model',                      //Column 0
          'TensorFlow Lite \n Coral',
          'TensorFlow Lite EdgeTPU \n Coral'],        //Column 1
          ['InceptionV1', 73, 32], //row 1
          ['InceptionV2', 72, 37], //row 2
          ['InceptionV3', 74, 14], //row 3
          ['InceptionV4', 0, 5], //row 4
          ['MobileNetV2',  0, 34], //row 5
          ['MobileNetV2 + SSD',  0, 45]  //row 6
        ]);
        var Coral_materialOptions_cpu = {
          width: 900,
          chart: {
            title: 'Model Vs CPU Load per backend',
          },
          series: {
          },
          axes: {
            y: {
              distance: {side: 'left',label: 'CPU Load'}, // Left y-axis.
            }
          }
        };
        var materialChart_coral_cpu = new google.charts.Bar(chartDiv_Cpu_Coral);
        view_coral_cpu = new google.visualization.DataView(table_models_cpu_coral);
        function drawMaterialChart() {
          var materialChart_coral_cpu = new google.charts.Bar(chartDiv_Cpu_Coral);
          materialChart_coral_cpu.draw(table_models_cpu_coral, google.charts.Bar.convertOptions(Coral_materialOptions_cpu));
          init_charts();
        }
        function init_charts(){
          view_coral_cpu.setColumns([0,1, 2]);
          materialChart_coral_cpu.draw(view_coral_cpu, Coral_materialOptions_cpu);
        }
         drawMaterialChart();
         drawMaterialChart();
    };
        }
 
</script>
</script>
</html>
</html>
<noinclude>
<noinclude>
{{GstInference/Foot|Example Applications/DispTec|Model Zoo}}
{{GstInference/Foot|Example Applications/DispTec|Model Zoo}}
</noinclude>
</noinclude>

Latest revision as of 20:39, 4 September 2024




Previous: Example Applications/DispTec Index Next: Model Zoo





GstInference Benchmarks - Introduction

This wiki summarizes a series of benchmarks on different hardware platforms based on the run_benchmark.sh bash script that can be found in the official GstInference repository. The script is based on the following GStreamer pipeline:

#Script to run each model
run_all_models(){

  model_array=(inceptionv1 inceptionv2 inceptionv3 inceptionv4 tinyyolov2 tinyyolov3)
  model_upper_array=(InceptionV1 InceptionV2 InceptionV3 InceptionV4 TinyYoloV2 TinyYoloV3)
  input_array=(input input input input input/Placeholder inputs )
  output_array=(InceptionV1/Logits/Predictions/Reshape_1 Softmax InceptionV3/Predictions/Reshape_1 
  InceptionV4/Logits/Predictions add_8 output_boxes )

  mkdir -p logs/
  rm -f logs/*

  for ((i=0;i<${#model_array[@]};++i)); do
    echo Perf ${model_array[i]}
    gst-launch-1.0 \
    filesrc location=$VIDEO_PATH num-buffers=600 ! decodebin ! videoconvert ! \
    perf print-arm-load=true name=inputperf ! tee name=t t. ! videoscale ! queue ! net.sink_model t. ! queue ! net.sink_bypass \
    ${model_array[i]} backend=$BACKEND name=net backend::input-layer=${input_array[i]} backend::output-layer=${output_array[i]} \
    model-location="${MODELS_PATH}${model_upper_array[i]}_${INTERNAL_PATH}/graph_${model_array[i]}${EXTENSION}" \
    net.src_bypass ! perf print-arm-load=true name=outputperf ! videoconvert ! fakesink sync=false > logs/${model_array[i]}.log
  done
}

Test benchmark video

The following video was used to perform the benchmark tests.
To download the video press right-click on the video and select 'Save video as' and save this on your computer.

Video 1. Test benchmark video

x86

The Desktop PC had the following specifications:

  • Intel(R) Core(TM) Core i7-7700HQ CPU @ 2.80GHz
  • 12 GB RAM
  • Linux 4.15.0-106-generic x86_64 (Ubuntu 16.04)
  • GStreamer 1.8.3

FPS Measurements

CPU Load Measurements

Jetson AGX Xavier

The Jetson Xavier power modes used were 2 and 6 (more information: Supported Modes and Power Efficiency)

  • View current power mode:
$ sudo /usr/sbin/nvpmodel -q
  • Change current power mode:
sudo /usr/sbin/nvpmodel -m x

Where x is the power mode ID (e.g. 0, 1, 2, 3, 4, 5, 6).

FPS Measurements

CPU Load Measurements

Jetson TX2

FPS Measurements

CPU Load Measurements

Jetson Nano

FPS Measurements

CPU Load Measurements

Google Coral

The following benchmarks were performed on the Coral Dev Board.

FPS Measurements

CPU Load Measurements


Previous: Example Applications/DispTec Index Next: Model Zoo