Coder Social home page Coder Social logo

xiaoxiong-lan / rgbt-tracking Goto Github PK

View Code? Open in Web Editor NEW

This project forked from xuboyue1999/rgbt-tracking

0.0 0.0 0.0 95.04 MB

This repository contains the necessary tools for RGBT tracking, including datasets(GTOT, RGBT234, LasHeR), evaluation tools, visualization tools, and results of existing works.

Python 5.13% MATLAB 94.87%

rgbt-tracking's Introduction

🌟New Function🌟 Now, you can plot the error curve for a single sequence, with different colors representing different trackers. The Y-axis represents tracking error, and the X-axis represents frame count. The plotting method can be seen in draw.py. image

RGBT-Tracking

This repository contains the necessary tools for RGBT tracking, including datasets(GTOT, RGBT234, LasHeR), evaluation tools, visualization tools, and results of existing works.

☘️Datasets☘️

🌟GTOT🌟

This dataset is derived from the paper "Learning Collaborative Sparse Representation for Grayscale-thermal Tracking" published in 2016 IEEE Transactions on Image Processing (T-IP) by Li et al. from the research group led by Professor Chenglong Li at Anhui University.

The dataset comprises 50 video pairs consisting of sequences of grayscale and thermal infrared images, with each pair having statistical bias. Additionally, the dataset includes ground truth annotations for each frame of the videos and two evaluation metrics.

For more information about the dataset and to download it, please visit gtot.md

image

🌟RGBT234🌟

This dataset is from the paper "RGB-T Object Tracking: Benchmark and Baseline" published in 2019 Pattern Recognition (PR) by Li et al. from the research group led by Professor Chenglong Li at Anhui University.

Key features of the dataset include:

The dataset comprises 234 pairs of RGB-T video sequences along with their corresponding ground truth annotations (GroundTruth). The video sequences are labeled with 12 attributes. The total number of frames in the dataset is 234,000, with the longest video sequence containing 8,000 frames.

For more information about the dataset and to download it, please visit rgbt234.md

image

🌟LasHeR🌟

LasHeR consists of 1224 visible and thermal infrared video pairs with more than 730K frame pairs in total. Each frame pair is spatially aligned and manually annotated with a bounding box, making the dataset well and densely annotated. LasHeR is highly diverse capturing from a broad range of object categories, camera viewpoints, scene complexities and environmental factors across seasons, weathers, day and night. Induced by real-world applications, several new challenges are take into consideration in data creation.

For more information about the dataset and to download it, please visit lasher.md

image

☘️Evaluation☘️

The primary evaluation metrics for RGBT single-object tracking include Precision Rate(PR), Success Rate (SR), and Normalized Precision Rate(NPR).

PR is a measure that describes the Euclidean distance between the center points of the predicted bounding box and the ground truth bounding box. It is generally considered that if the Euclidean distance between the center points of the ground truth bounding box and the predicted bounding box is less than 20 pixels, it indicates accurate localization. However, in the GTOT dataset, this threshold is set to 5 pixels.

The formula for PR is PR = NP/NT,

where NP represents the number of frames where the distance between the tracked algorithm's located target center point and the true target center point is less than a certain threshold, and NT represents the total number of frames.

SR is calculated by measuring the intersection-over-union (IoU) of pixels within the predicted bounding box and the ground truth bounding box. If the IoU exceeds a certain threshold, it is considered a successful tracking.

The formula for SR is SR = NS/NT,

where NS represents the number of frames in which the tracking algorithm successfully tracked the target, and NT represents the total number of frames.

NPR, which stands for Normalized Precision, involves normalizing the predicted bounding box (bbox) and ground truth bbox. By doing so, the performance of the model can be assessed without considering the size of the image resolution.

🌟GTOT🌟

We have developed a MATLAB-based evaluation toolbox for GTOT, which includes

  • methods for calculating PR and SR metrics, generating metric curves, and comparing with existing trackers
  • the original tracking result files for the majority of existing trackers
  • visualization and analysis methods for tracking
  • etc

For detailed instructions on how to use the toolbox and for more information, please refer to the documentation available at GTOT-toolkit. It will provide comprehensive guidance on utilizing the toolbox and accessing additional details about its functionality.

image

🏆conparision🏆

Here are the metrics of existing trackers on the GTOT dataset: image

The sources of the trackers are indicated in the following table:

Tracker Paper year Source
MTNet MTNet: Learning Modality-aware Representation with Transformer for RGBT Tracking 2023 ICME
HMFT Visible-thermal uav tracking: A large-scale benchmark and new baseline 2022 CVPR
MIRNet Mirnet: A robust rgbt tracking jointly with multi-modal interaction and refinement 2022 ICME
APFNet Attribute-based progressive fusion network for rgbt tracking 2022 AAAI
DMCNet Duality-gated mutual condition network for rgbt tracking 2022 IEEE TNNLS
SiamCDA Siamcda: Complementarityand distractor-aware rgb-t tracking based on siamese network 2022 TCSVT
JMMAC Jointly modeling motion and appearance cues for robust rgb-t tracking 2021 IEEE Transactions on Image Processing
ADRNet Learning Adaptive Attribute-Driven Representation for Real-Time RGB-T Tracking 2020 IJCV
CMPP Cross-modal pattern-propagation for rgb-t tracking 2020 CVPR
CAT Challenge-aware RGBT tracking 2020 ECCV
MacNet Object tracking in RGB-T videos using modal-aware attention network and competitive learning 2020 Sensors
MANet Multi-adapter RGBT tracking 2019 ICCVW

🌟RGBT234🌟

results are provided 🚧building🚧

🌟LasHeR🌟

The official repository for the Lasher dataset provides evaluation tools.

Based on this, we have added some metrics and result files for the latest trackers.

  • methods for calculating PR、SR and NPR metrics, generating metric curves, and comparing with existing trackers
  • the original tracking result files for the majority of existing trackers
  • etc

For detailed instructions on how to use the toolbox and for more information, please refer to the documentation available at Lasher-toolkit. It will provide comprehensive guidance on utilizing the toolbox and accessing additional details about its functionality.

🏆comparision🏆

Here are the metrics of existing trackers on the LasHeR dataset: image

The sources of the trackers are indicated in the following table:

Tracker Paper year Source
MTNet MTNet: Learning Modality-aware Representation with Transformer for RGBT Tracking 2023 ICME
MIRNet Mirnet: A robust rgbt tracking jointly with multi-modal interaction and refinement 2022 ICME
APFNet Attribute-based progressive fusion network for rgbt tracking 2022 AAAI
DMCNet Duality-gated mutual condition network for rgbt tracking 2022 IEEE TNNLS
FANet Quality-aware feature aggregation network for robust rgbt tracking 2021 IEEE Transactions on Intelligent Vehicles
MacNet Object tracking in RGB-T videos using modal-aware attention network and competitive learning 2020 Sensors
MANet++ Object tracking in rgb-t videos using modal-aware attention network and competitive learning 2020 Sensors
CAT Challenge-aware RGBT tracking 2020 ECCV
MANet Multi-adapter RGBT tracking 2019 ICCVW
DAFNet Deep adaptive fusion network for high performance rgbt tracking 2019 ICCVW
mfdimp Multi-modal fusion for end-to-end rgb-t tracking 2019 ICCVW
DAPNet Dense feature aggregation and pruning for rgbt tracking 2019 MM
SGT++ Rgb-t object tracking: benchmark and baseline 2019 Pattern Recognization
CMR Cross modal ranking with soft consistency and noisy labels for robust rgb-t tracking 2018 ECCV
SGT Weighted sparse representation regularized graph learning for rgbt object tracking 2017 MM

rgbt-tracking's People

Contributors

xuboyue1999 avatar

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.