GstInference Releases
(Redirected from GstInference/Releases)
Make sure you also check GstInference's companion project: R2Inference  | 
| GstInference | 
|---|
| Introduction | 
| Getting started | 
| Supported architectures | 
| 
 InceptionV1 InceptionV3 YoloV2 AlexNet | 
| Supported backends | 
| 
 Caffe  | 
| Metadata and Signals | 
| Overlay Elements | 
| Utils Elements | 
| Legacy pipelines | 
| Example pipelines | 
| Example applications | 
| Benchmarks | 
| Model Zoo | 
| Project Status | 
| Contact Us | 
   | 
This page summarizes all GstInference and R2Inference releases. Check GitHub for more info!
v0.1.0
Introduced features:
- Video inference base class
 - Support for the following architectures:
- GoogLeNet (Inception v4)
 - Tiny Yolo v2
 
 - Support for the following backends:
- NCSDK
 - TensorFlow
 
 - Support for the following platforms:
- Intel Movidius Neural Compute Stick (version 1)
 - NVIDIA Jetson AGX Xavier
 - x86 systems
 
 - Set backend from GStreamer
 - Set backend properties from GStreamer
 - Set model location from GStreamer
 - Support play-stop-play transitions on GstInference elements
 
Known issues:
- NCSDK does not support multiple calls to the inference engine from the same thread. This causes transitions when using the NCSDK backend to fail after the second play.
 - Changing the backend on a stoped pipeline will fail with a segmentation fault.
 - When using the TensorFlow backend compiled with GPU support, the pipeline will sometimes fail to start.
 - The plugins throw segmentation fault when using nvarguscamerasrc as input on Xavier.
 
v0.2.0
Introduced features:
- Stability improved.
 - Inference on InceptionV4 improved.
 - Metadata structure created to save the inference result.
 - Metadata is attached to model buffer and bypass buffer (if available).
 - Signal created to use the inference result at the application level.
 - The inference result is not displayed by the inference element unless the log level is set to 6.
 - Pads control improved, bypass pad is now optional.
 - GoogleNet renamed to InceptionV4.
 - TinyYolo renamed to TinyYoloV2.
 - Jetson TX2 added to supported architectures.
 - Examples section improved to smaller pipelines.
 - Overlay plugins added:
- GstClassificationOverlay for InceptionV4 metadata.
 - GstDetectionOverlay for TinyYoloV2 metadata.
 
 - Example applications were added.
 - The segmentation fault with NVarguscamerasrc is fixed.
 - TensorFlow is able to run on CPU and GPU.
 
Known issues
- NCSDK does not support multiple calls to the inference engine from the same thread. This causes transitions when using the NCSDK backend to fail after the second play.
 - Changing the backend on a stoped pipeline will fail with a segmentation fault.
 
v0.3.0
Introduced features:
- Fixed bug with gtk-doc compilation on GStreamer 1.8
 - Support for the following architectures:
- Inception v3
 - Inception v2
 - Inception v1
 
 
Known issues
- NCSDK does not support multiple calls to the inference engine from the same thread. This causes transitions when using the NCSDK backend to fail after the second play.
 - Changing the backend on a stoped pipeline will fail with a segmentation fault.
 
v0.4.0
Introduced features:
- Support on Inception and TinyYOLO elements for the following image formats:
- RGBx
 - RGBA
 - xRGB
 - ARGB
 - BGR
 - BGRx
 - BGRA
 - xBGR
 - ABGR
 
 - Support for FaceNet architecture
 - Added Embedding overlay plugin for face detection visualization
 
Known issues
- NCSDK does not support multiple calls to the inference engine from the same thread. This causes transitions when using the NCSDK backend to fail after the second play.
 - Changing the backend on a stoped pipeline will fail with a segmentation fault.
 
v0.5.0
Introduced features:
- Support for the following architectures:
- ResNet50V1
 - Tiny Yolo V3
 - FaceNetV1
 - MobileNetV2
 
 
- Support for the following backends:
- NCSDK
 - TensorFlow
 
 
- Support for the following platforms:
 - Intel Movidius Neural Compute Stick (version 1)
 - NVIDIA Jetson AGX Xavier
 - x86 systems
 - NVIDIA TX2
 - i.MX8
 
Known issues:
- NCSDK does not support multiple calls to the inference engine from the same thread. This causes the NCSDK backend to fail after the second start.
 - Changing the backend on a stoped pipeline will fail with a segmentation fault.
 
v0.6.0
Introduced features:
- Improved ResNet50V1 inference result
 - Preprocess and Postprocess factorized as general files.
 - Debug information factorized as a general file.
 - Tests added.
 - Improved internal caps negotiation on bypass pad.
 - Copyright license LGPL added.
 
Supported platforms:
- Intel Movidius Neural Compute Stick (version 1)
 - NVIDIA Jetson AGX Xavier
 - NVIDIA TX2
 - NVIDIA Nano
 - x86 systems
 - i.MX8
 
Known issues:
- NCSDK does not support multiple calls to the inference engine from the same thread. This causes the NCSDK backend to fail after the second start.
 
v0.7.0
Introduced features:
- Pkg-config support
 - License update to LGPL
 - New inference meta hierarchical approach
 - TFlite backend support
 - New elements using new meta:
- Detectioncrop
 - Inferencefilter
 - Inferencedebug
 - Inferenceoverlay
 
 - Bug fixes
 
v0.7.1
Introduced features:
- Revert hotfix in dev-0.6 that modify the number of predictions.
 
v0.8.0
Introduced features:
- Add new inferenceutils plugin to absorb inferencecrop, inferencefilter and inferencedebug.
 - Show root prediction on inferenceoverlay.
 - Fix prediction size test.
 - Fix tinyyolo3 postprocess to use new meta.
 - Bug fixes
 
v0.9.0
Introduced features:
- Fix Yolo probabilities in inferencemeta
 - Support for OpenCV 4
 - Support for doubles in backend properties
 - New inferencebin helper element
 
	