Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.
Tip

Video stream and scene requirements for the Human pose detector and its sub-detectors

Image requirements for the Human pose detector and its sub-detectors

Hardware requirements for neural analytics operation

Optimizing the operation of neural analytics on GPU in Windows OS

Optimizing the operation of neural analytics on GPU in Linux OS

To configure the Human pose detector, do the following:

  1. Go to the
  2.  Detection Tools
  3. Detectors tab.
  4. Below the required camera, click Create…  Category: Poses Human pose detector.

By default, the detection tool detector is enabled and set to detect poses.

If necessary, you can change the detection tool detector parameters. The list of parameters is given in the table:

ParameterValueDescription
Object features
Record objects trackingYesThe metadata of the video stream is recorded to the database by default. To disable the parameter, select the No value
No
Video streamMain streamIf the camera supports multistreaming, select the stream for which detection is needed. Selecting a low-quality video stream allows reducing the load on the
Server
server.
Second stream
Other
EnableYesThe
detection tool
detector is enabled by default. To disable
the detection tool
, select the No value
No
NameHuman pose detectorEnter the
detection tool
detector name or leave the default name
Decoder modeAutoSelect a processing resource for decoding video streams. When you select a GPU, a stand-alone graphics card takes priority (when decoding with NVIDIA NVDEC chips). If there is no appropriate GPU, the decoding will use the Intel Quick Sync Video technology. Otherwise, CPU resources
will be
are used for decoding
CPU
GPU
HuaweiNPU
Number of frames processed per second3

Specify the number of frames that the

detection tool

detector will process per second. The value must be in the

range 

range [0.016; 100]

Note
titleAttention!

With static persons in the scene, set the FPS to no less than 2. With moving persons in the scene, the FPS must be at least 4

The higher the FPS value, the higher the accuracy of pose detection, but the load on the selected processor is higher as well. For FPS

=

equal to 1, the accuracy

will be

is no less than 70%.

This parameter varies depending on the object speed of movement. To solve typical tasks, an FPS value from 3 to 20 is sufficient. Examples:

  • pose detection for moderately moving objects (without sudden movements)—FPS 3;
  • pose detection for moving objects—FPS 12.


TypeHuman pose detectorName of the
detection tool
detector type (non-editable field)

Advanced settings

If detection of small objects or objects in areas far away from the camera is ineffective, you can use the scanning mode (see Configuring the scanning mode).

 This

This mode doesn’t provide absolute detection accuracy

,

but can improve detection performance

Neural network file
Select

Specify a neural network file. The standard neural networks for different processor types are located in the C:\Program Files\Common Files\AxxonSoft\DetectorPack\NeuroSDK directory.

 You

You don't need to select the standard neural networks in this field, the system will automatically select the required one. If you use a custom neural network, enter a path to the file.

info
Note
title
NoteFor the Note
titleAttention!
Attention!
  • You cannot specify the network file in Windows OS. You must place the neural network file locally, that is, on the same server where you install Axxon One.
  • For correct neural network operation on Linux OS, place the corresponding file locally in the /opt/AxxonSoft/DetectorPack/NeuroSDK directory
  • or in the network folder with the corresponding access rights.
Scanning windowYes
 The
The scanning mode is disabled by default. To enable the scanning mode, select
the
the Yes value
No
Scanning window height540

The height and width of the scanning window are determined according to the actual size of the frame and the required number of windows. For example, the real frame size is 1920×1080 pixels. To divide the frame into four equal windows, set the width of the scanning window to 960 pixels and the height to 540 pixels

Scanning window step height540

The scanning step determines the relative offset of the windows. If the step is equal to the height and width of the scanning window, respectively, the segments will line up one after another. Reducing the height or width of the scanning step will increase the number of windows due to their overlapping each other with an offset. This will increase the detection accuracy

,

but will also increase the CPU load

Note
titleAttention!

The height and width of the scanning step must not be greater than the height and width of the scanning window. The

detection tool

detector won't operate with such settings.


Scanning window step width480

The scanning step determines the relative offset of the windows. If the step is equal to the height and width of the scanning window, respectively, the segments will line up one after another. Reducing the height or width of the scanning step will increase the number of windows due to their overlapping each other with an offset. This will increase the detection accuracy

,

but will also increase the CPU load

Note
titleAttention!

The height and width of the scanning step must not be greater than the height and width of the scanning window. The

detection tool

detector won't operate with such settings.

Scanning window width480The height and width of the scanning window are determined according to the actual size of the frame and the required number of windows. For example, the real frame size is 1920×1080 pixels. To divide the frame into four equal windows, set the width of the scanning window to 960 pixels and the height to 540 pixels
Track lifespan (starting with Detector Pack 3.14)
Yes

By default, the parameter is disabled. If you want to display the track lifespan for an object in seconds, select the Yes value

 

No
Basic settings
Mode







CPU

Select the operation mode of the

detection tool and

detector and neural filter (see Hardware requirements for neural analytics operationSelecting Nvidia GPU when configuring detectors). Neural filter operation mode is used only if you select the Yes value in the Neural filter parameter.

Note
titleAttention!
other
  • another processing resource than the CPU, this device
will carry the
  • carries most of the computing load.
 However
  • However, the CPU
will
  • is also
be
  • used to run the
detection tool
  • detector.
  • For the Person Down Detection and Sitting Person Detection, the accuracy can depend on the selected processor. If the recognition quality deteriorates when you change the processor, we recommend selecting the best values of the detector parameters
of detection tools
  • and configuring perspective empirically (see Sitting person detector, Person down detector).
  • Starting with Detector Pack 3.11, Intel HDDL and Intel NCS aren’t supported.
  • Starting with Detector Pack 3.14, Intel Multi-GPU and Intel GPU 0-3 are supported.




 

 

 

 

 

 

 

Nvidia GPU 0
Nvidia GPU 1
Nvidia GPU 2
Nvidia GPU 3
Intel NCS (not supported)
Intel HDDL (not supported)
Intel GPU
Intel Multi-GPU
Intel GPU 0
Intel GPU 1
Intel GPU 2
Intel GPU 3
Huawei NPU
Pose estimation neural network (starting with Detector Pack 3.14)



Pose estimation

Select a pose estimation neural network. The names may indicate the size of the neural network. The size (Nano, Middle, Large) of the neural network indicates the amount of resources consumed. The larger the neural network, the higher the accuracy but the higher the loadon the processor

 

 

Pose estimation (Nano)
Pose estimation (Middle)
Pose estimation (Large)
Neural network filter
Neural filter

YesNeural filter is disabled by default. To use the neural filter to filter out parts of tracks, select the Yes value
No
Neural filter file
Select a neural network file

By default, the entire frame is a detection area. In the preview window, you can specify the custom detection areas and skip areas. To specify areas, right-click the frame , and select the required area (see see Configuring a skip areaConfiguring a detection area). 

Image RemovedImage Added

The following logic is used for this:

  • if you specify only detection areas, no detection

...

  • is performed in all other parts of the frame;
  • if you specify only skip areas, detection

...

  • is performed in all other parts of the frame.
    Image Modified
Info
titleNote
  • For convenience of configuration, you can "freeze" the frame. Click the Image Modified button. To cancel the action, click this button again.
  • The detection area is displayed by default. To hide it, click the Image Modified button. To cancel the action, click this button again.
  • To delete the selected area, click the Image Modified button. 

To save the parameters of the detection tooldetector, click the Apply Image Modified button. To cancel the changes, click the Cancel Image Modified button.

Configuring the Human pose detector is complete. If necessary, you can create and configure the required sub-detectors on the basis of the Human pose detector (see Configuring the Human pose detector sub-detectors).