Publisher
Springer Nature Switzerland
Reference14 articles.
1. Abdar, M., et al.: A review of deep learning for video captioning (2023). https://doi.org/10.48550/ARXIV.2304.11431
2. Chen, D.L., Dolan, W.B.: Collecting highly parallel data for paraphrase evaluation. In: Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics (ACL-2011), Portland, OR (2011)
3. Dilawari, A., Khan, M.U.G., Farooq, A., Rehman, Z.U., Rho, S., Mehmood, I.: Natural language description of video streams using task-specific feature encoding. IEEE Access 6, 16639–16645 (2018)
4. Fonseca, C.M., Paiva, J.G.S.: A system for visual analysis of objects behavior in surveillance videos. In: 2021 34th SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI), pp. 176–183. IEEE (2021)
5. Iashin, V., Rahtu, E.: Multi-modal dense video captioning. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops (2020)