MaochengHu 576cda45b8 first commit | há 2 anos atrás | |
---|---|---|
.. | ||
cpp | há 2 anos atrás | |
python | há 2 anos atrás | |
README.md | há 2 anos atrás | |
README_cn.md | há 2 anos atrás | |
README_en.md | há 2 anos atrás |
English | 简体中文
PP-Tracking is the first open source real-time Multi-Object Tracking system, and it is based on PaddlePaddle deep learning framework. It has rich models, wide application and high efficiency deployment.
PP-Tracking supports two paradigms: single camera tracking (MOT) and multi-camera tracking (MTMCT). Aiming at the difficulties and pain points of actual business, PP-Tracking provides various MOT functions and applications such as pedestrian tracking, vehicle tracking, multi-class tracking, small object tracking, traffic statistics and multi-camera tracking. The deployment method supports API and GUI visual interface, and the deployment language supports Python and C++, The deployment platform environment supports Linux, NVIDIA Jetson, etc.
PP-tracking provides AI studio public project cases. Please refer to this tutorial.
PP-Tracking supports Python predict and deployment. Please refer to this doc.
PP-Tracking supports C++ predict and deployment. Please refer to this doc.
PP-Tracking supports GUI predict and deployment. Please refer to this doc.
PP-Tracking supports two paradigms: single camera tracking (MOT) and multi-camera tracking (MTMCT).
PP-Tracking provids multi-scenario pre-training models and the exported models for deployment:
Scene | Dataset | MOTA | Speed(FPS) | config | model weights | inference model |
---|---|---|---|---|---|---|
pedestrian | MOT17 | 65.3 | 23.9 | config | download | download |
pedestrian(small objects) | VisDrone-pedestrian | 40.5 | 8.35 | config | download | download |
vehicle | BDD100k-vehicle | 32.6 | 24.3 | config | download | download |
vehicle(small objects) | VisDrone-vehicle | 39.8 | 22.8 | config | download | download |
multi-class | BDD100k | - | 12.5 | config | download | download |
multi-class(small objects) | VisDrone | 20.4 | 6.74 | config | download | download |
Note:
model weights
means the weights saved directly after PaddleDetection training. For more tracking model weights, please refer to modelzoo, you can also train according to the corresponding model config file and get the model weights.inference model
means the model weights with only forward parameters after exported, because only forward parameters are required during the deployment of PP-Tracking project. It can be downloaded and exported according to modelzoo, you can also train according to the corresponding model config file and get the model weights, and then export them。In exported model files, there should be infer_cfg.yml
,model.pdiparams
,model.pdiparams.info
and model.pdmodel
four files in total, which are generally packaged in tar format.@ARTICLE{9573394,
author={Zhu, Pengfei and Wen, Longyin and Du, Dawei and Bian, Xiao and Fan, Heng and Hu, Qinghua and Ling, Haibin},
journal={IEEE Transactions on Pattern Analysis and Machine Intelligence},
title={Detection and Tracking Meet Drones Challenge},
year={2021},
volume={},
number={},
pages={1-1},
doi={10.1109/TPAMI.2021.3119563}
}
@InProceedings{bdd100k,
author = {Yu, Fisher and Chen, Haofeng and Wang, Xin and Xian, Wenqi and Chen,
Yingying and Liu, Fangchen and Madhavan, Vashisht and Darrell, Trevor},
title = {BDD100K: A Diverse Driving Dataset for Heterogeneous Multitask Learning},
booktitle = {IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
month = {June},
year = {2020}
}
@article{zhang2020fair,
title={FairMOT: On the Fairness of Detection and Re-Identification in Multiple Object Tracking},
author={Zhang, Yifu and Wang, Chunyu and Wang, Xinggang and Zeng, Wenjun and Liu, Wenyu},
journal={arXiv preprint arXiv:2004.01888},
year={2020}
}
@inproceedings{Wojke2018deep,
title={Deep Cosine Metric Learning for Person Re-identification},
author={Wojke, Nicolai and Bewley, Alex},
booktitle={2018 IEEE Winter Conference on Applications of Computer Vision (WACV)},
year={2018},
pages={748--756},
organization={IEEE},
doi={10.1109/WACV.2018.00087}
}