论文标题
用于放大车牌识别和统一数据集的分解生成网络
Disentangled Generation Network for Enlarged License Plate Recognition and A Unified Dataset
论文作者
论文摘要
车牌识别在许多实际应用中起着至关重要的作用,但是由于低分辨率,污染,低照明和遮挡的因素,很难识别大型车辆的车牌。为了克服上述因素,运输管理部通常会引入车辆后部后面的放大车牌。但是,放大的车牌具有高度的多样性,因为它们的位置,大小和样式都不是标准化。此外,背景区域包含各种嘈杂的信息,这些信息极大地打扰了车牌字符的识别。现有作品尚未研究这个具有挑战性的问题。在这项工作中,我们首先解决了放大的车牌识别问题,并贡献了包含9342张图像的数据集,该数据集涵盖了真实场景的大多数挑战。但是,创建的数据仍然不足以训练扩大的车牌识别的深度方法,并且构建大规模培训数据非常耗时和劳动力成本高。为了解决这个问题,我们提出了一个基于解开的生成网络(DGNET)的新型任务级解剖生成框架,该框架以端到端的方式将一代置于文本生成和背景生成中,以有效地确保多样性和完整性,以确保强大的放大牌照牌照识别。进行了创建数据集的广泛实验,我们在三个代表性文本识别框架中证明了所提出方法的有效性。
License plate recognition plays a critical role in many practical applications, but license plates of large vehicles are difficult to be recognized due to the factors of low resolution, contamination, low illumination, and occlusion, to name a few. To overcome the above factors, the transportation management department generally introduces the enlarged license plate behind the rear of a vehicle. However, enlarged license plates have high diversity as they are non-standard in position, size, and style. Furthermore, the background regions contain a variety of noisy information which greatly disturbs the recognition of license plate characters. Existing works have not studied this challenging problem. In this work, we first address the enlarged license plate recognition problem and contribute a dataset containing 9342 images, which cover most of the challenges of real scenes. However, the created data are still insufficient to train deep methods of enlarged license plate recognition, and building large-scale training data is very time-consuming and high labor cost. To handle this problem, we propose a novel task-level disentanglement generation framework based on the Disentangled Generation Network (DGNet), which disentangles the generation into the text generation and background generation in an end-to-end manner to effectively ensure diversity and integrity, for robust enlarged license plate recognition. Extensive experiments on the created dataset are conducted, and we demonstrate the effectiveness of the proposed approach in three representative text recognition frameworks.