Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

...

Tip

Video stream and scene requirements for neural counter operation

Hardware requirements for neural analytics operation

To configure Neurocounter, do the following:

  1. To record mask (highlighting of the recognized objects) to the archive, select Yes for the corresponding parameter (1).
    Image Modified
  2. If the camera supports multistreaming, select the stream for which detection is needed (2). 
    Image Modified
  3. Select a processing resource for decoding video streams (3). When you select a GPU, a stand-alone graphics card takes priority (when decoding with NVIDIA NVDEC chips). If there is no appropriate GPU, the decoding will use the Intel Quick Sync Video technology. Otherwise, CPU resources will be used for decoding.
  4. If you need to outline the recognized objects in the preview window, select Yes for the Detected objects parameter (4).
  5. Set the recognition threshold for the objects in percent (5). If the recognition probability falls below the specified value, the data will be ignored. The higher the value, the higher the recognition accuracy, but some triggers may not be considered.
  6. Set the frame rate value for the detection tool to process per second (6). This value should be in the range [0.,016; 100]. 

    Info
    titleNote

    The default values (3 output frames and 1 FPS) indicate that Neurocounter will analyze one frame every second. If Neurocounter detects the specified number of objects (or more) on 3 frames, then it triggers.


  7. Select the processor for the neural network CPU, one of NVIDIA GPUs , or one of Intel CPUs GPUs (7, see Hardware requirements for neural analytics operation, General Information on Configuring Detection). 

    Note
    titleAttention!

    If you specify other processing resource than the CPU, this device will carry the most of computing load. However, the CPU will also be used to run Neurocounter.


    Note
    titleAttention!

    It may take several minutes to launch the algorithm on NVIDIA GPU NVIDIA GPU after you apply the settings. You can use caching to speed up future launches (see Configuring the acceleration of GPU-based neuroanalytics).


  8. In the Object type field  field (11), select the object type for counting, or in the the Neural network file field (8), select the neural network file.

    Note
    titleAttention!

    To train your neural network, contact AxxonSoft (see  Data collection requirements for neural network training).

    A trained neural network for a particular scene allows you to detect only objects of a certain type (e.g. person, cyclist, motorcyclist, etc.). 

    If the neural network file is not specified, the default file will be used, which is selected depending on the selected object type (11) and the selected processor for the neural network operation (7).


    Info
    titleNote

    For correct neural network operation on Linux, place the corresponding file in the /opt/AxxonSoft/DetectorPack/NeuroSDK directory.


  9. Set the triggering condition for the neurocounterNeurocounter:

    1. In the Number of alarm objects field (9), set the threshold value for the number of objects in the FOV. 

    2. In the Trigger upon count field (12), select when you want to generate the trigger when the number of objects in the detection area is greater or less than the threshold value. 

  10. Set the minimum number of frames on which Neurocounter should detect objects in order to trigger (10). The value should be in the range [2; 20].

  11. In the preview window, you can set the detection areas with the help of anchor points much like privacy masks in Scene Analytics detection tools (see Setting General Zones for Scene Analytics analytics detection tools). By default, the entire FOV is a detection area.
    Image Modified
  12. Click the Apply button.

It is possible to display the sensor and the number of objects in the controlled area in the video surveillance window on the layout. To configure this option, do the following:

  1. Switch to the Layout Editing mode (see Switching to layout editing mode).
  2. Place the sensor anywhere in the FOV.
    Image Modified
  3. Customize the font. To do this, click the Image Removed Image Added button.
  4. Save the layout (see Exiting layout editing mode). As a result, the sensor and the number of objects will be displayed in the selected place:
    Image Modified