|
[1] G. Praetorius, “Vessel Traffic Service (VTS): A maritime information service or traffic control system?: Understanding everyday performance and resilience in a socio-technical system under change,” Ph.D. Dissertation, Dept. Shipping and Marine Tech., Chalmers tekniska högskola, Gothenburg, Sweden, 2014. [2] A. Harati-Mokhtari et al., “Automatic Identification System (AIS): Data reliability and human error implications,” The Journal of Navigation, vol. 60, no. 3, pp. 373-389, 2007. [3] S. Ren et al., “Faster R-CNN: Towards real-time object detection with region proposal networks,” Advances in Neural Information Processing Systems, vol. 28, 2015. [4] W. Liu et al., “SSD: Single shot multibox detector,” in Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 2016, pp. 21-37. [5] J. Redmon and A. Farhadi, “Yolov3: An incremental improvement,” April 2018. [Online]. Available: https://arxiv.org/abs/1804.02767. [6] A. Bochkovskiy et al., “Yolov4: Optimal speed and accuracy of object detection,” April 2020. [Online]. Available: https://arxiv.org/abs/2004.10934 [7] A. Bewley et al., “Simple online and realtime tracking,” in Proceedings of the IEEE International Conference on Image Processing, Phoenix, Arizona, 2016, pp. 3464-3468. [8] N. Wojke, A. Bewley, and D. Paulus, “Simple online and realtime tracking with a deep association metric,” in Proceedings of the IEEE International Conference on Image Processing, Beijing, China, 2017, pp. 3645-3649. [9] Z. Wang et al., “Towards real-time multi-object tracking,” in Proceedings of the European Conference on Computer Vision, Glasgow, United Kingdom, 2020, pp. 107-122. [10] T. Y. Lin et al., “Feature pyramid networks for object detection,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, Hawaii, 2017, pp. 2117-2125. [11] Y. Zhang et al., “Fairmot: On the fairness of detection and re-identification in multiple object tracking,” International Journal of Computer Vision, vol. 129, no. 11, pp. 3069-3087, 2021. [12] K. He et al., “Deep residual learning for image recognition,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, Nevada, 2016, pp. 770-778. [13] F. Yu et al., “Deep layer aggregation,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, Utah, 2018, pp. 2403-2412. [14] X. Zhou, V. Koltun, and P. Krähenbühl, “Tracking objects as points,” in Proceedings of the European Conference on Computer Vision, Glasgow, United Kingdom, 2020, pp. 474-490. [15] X. Zhou, D. Wang, and P. Krähenbühl, “Objects as points,” April 2019. [Online]. Available: https://arxiv.org/abs/1904.07850. [16] A. Newell, K. Yang, and J. Deng, “Stacked hourglass networks for human pose estimation,” in Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 2016, pp. 483-499. [17] A. Vaswani et al., “Attention is all you need,” Advances in Neural Information Processing Systems, vol. 30, 2017. [18] I. Sutskever, O. Vinyals, and Q. V. Le, “Sequence to sequence learning with neural networks,” Advances in Neural Information Processing Systems, vol. 27, 2014. [19] J. Alammar, “The Illustrated Transformer,” [Online]. Available: http://jalammar.github.io/illustrated-transformer/. [Accessed August 16, 2022]. [20] N. Carion et al., “End-to-end object detection with transformers,” in Proceedings of the European Conference on Computer Vision, Glasgow, United Kingdom, 2020, pp. 213-229. [21] X. Zhu et al., “Deformable detr: Deformable transformers for end-to-end object detection,” October 2020. [Online]. Available: https://arxiv.org/abs/2010.04159. [22] J. Dai et al., “Deformable convolutional networks,” in Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 2017, pp. 764-773. [23] P. Sun et al., “Transtrack: Multiple object tracking with transformer,” December 2020. [Online]. Available: https://arxiv.org/abs/2012.15460. [24] Y. Zhang et al., “Bytetrack: Multi-object tracking by associating every detection box,” October 2021. [Online]. Available: https://arxiv.org/abs/2110.06864. [25] T. Y. Lin et al., “Focal loss for dense object detection,” in Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 2017, pp. 2980-2988. [26] H. Rezatofighi et al., “Generalized intersection over union: A metric and a loss for bounding box regression,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, California, 2019, pp. 658-666. [27] A. Milan et al., “MOT16: A benchmark for multi-object tracking,” March 2016. [Online]. Available: https://arxiv.org/abs/1603.00831. [28] P. Dendorfer et al., “Mot20: A benchmark for multi object tracking in crowded scenes,” March 2020. [Online]. Available: https://arxiv.org/abs/2003.09003.
|