Why doesn't the "sensor fusion and tracking toolbox" have advanced tracking algorithms such as KCF, GOTURN, FairMOT, deepSort?

16 vues (au cours des 30 derniers jours)
KCF, GOTURN, FairMOT, deepSort are some of the best and successful algorithms in recent years, but unfortunately, why is there no such implementation in the "sensor fusion and tracking toolbox", "deeplearning toolbox" and "computer vision toolbox"?
  3 commentaires
cui,xingxing
cui,xingxing le 10 Juin 2021
@Walter Roberson Sorry, but what does this wiki link mean and is it relevant to the question?
Walter Roberson
Walter Roberson le 10 Juin 2021
I quote you,
"why is there no such implementation"
You wanted to know why. Now you know one of the reasons.

Connectez-vous pour commenter.

Réponse acceptée

Elad Kivelevitch
Elad Kivelevitch le 10 Juin 2021
Hi Cui,
To the best of my knowledge, the algorithms you cite are mostly used in vision-based tracking and they are intended for use on a single video stream, in other words a single-sensor system. Moreover, an algorithm like deepSORT is actually a combination of deep-learning detector and a SORT (simple, online, and real-time tracker), and if you dig deep into the SORT part, it's nothing more than a Hungarian association algorithm and Kalman-based filtering (from the link you shared). That combination of association and filtering is the same as trackerGNN in Sensor Fusion and Tracking, which means you just need a detector to feed it.
The algorithms we ship in Sensor Fusion and Tracking Toolbox are designed for multi-sensor, multi-modality tracking problems. Our goal is to enable tracking using a variety of sensors, including camera, but also radar, lidar, sonar, etc. Therefore, the trackers in Sensor Fusion and Tracking algorithms do not depend on any particular sensor type and the inputs to them are generic objectDetection objects.
You can easily create a vision-based tracking system if you combine the detectors that Computer Vision Toolbox offers with the trackers from Sensor Fusion and Tracking Toolbox. The Computer Vision Toolbox detector (and there are several types of them, some of them based on deep learning) will output bounding boxes and if you define objectDetection for each bounding box and you can use any tracking filter and tracker from Sensor Fusion and Tracking to complete the tracking. Please see:
Similarly, we use bounding boxes coming from Lidar Toolbox detectors to track objects. Some of these detectors are based on deep learning as well. There are several examples of that in Automated Driving Toolbox, Sensor Fusion and Tracking Toolbox, and Lidar Toolbox.
Thanks,
Elad
  1 commentaire
Walter Roberson
Walter Roberson le 10 Juin 2021
There is a school of thought that says that the role of a business should not be to give the customers what the customers actually need, but rather that the role of a business should be to give the customers what the customers think they need, and then to let the customers come to their own conclusions.
So in the spirit of that school of thought: Why has Mathworks not already released polished solutions by the end of the same day that the respective algorithms were published?

Connectez-vous pour commenter.

Plus de réponses (0)

Community Treasure Hunt

Find the treasures in MATLAB Central and discover how the community can help you!

Start Hunting!

Translated by