A New Transformer-Based Approach for Text Detection in Shaky and Non-shaky Day-Night Video

Publisher:
Springer
Publication Type:
Conference Proceeding
Citation:
Pattern Recognition, 2023, 14407 LNCS, pp. 30-44
Issue Date:
2023-01-01
Filename Description Size
978-3-031-47637-2_3.pdfPublished version4.89 MB
Adobe PDF
Full metadata record
Text detection in shaky and non-shaky videos is challenging because of variations caused by day and night videos. In addition, moving objects, vehicles, and humans in the video make the text detection problems more challenging in contrast to text detection in normal natural scene images. Motivated by the capacity of the transformer, we propose a new transformer-based approach for detecting text in both shaky and non-shaky day-night videos. To reduce the effect of object movement, poor quality, and other challenges mentioned above, the proposed work explores temporal frames for obtaining activation frames based on similarity and dissimilarity measures. For estimating similarity and dissimilarity, our method extracts luminance, contrast, and structural features. The activation frames are fed to the transformer which comprises an encoder, decoder, and feed-forward network for text detection in shaky and non-shaky day-night video. Since it is the first work, we create our own dataset for experimentation. To show the effectiveness of the proposed method, experiments are conducted on a standard dataset called the ICDAR-2015 video dataset. The results on our dataset and standard dataset show that the proposed model is superior to state-of-the-art methods in terms of recall, precision, and F-measure.
Please use this identifier to cite or link to this item: