用于几何信息学习的图结构运动分割方法
DOI:
CSTR:
作者:
作者单位:

1.福建农林大学机电工程学院;2.中国科学院海西研究院泉州装备制造研究中心;3.泉州职业技术大学

作者简介:

通讯作者:

中图分类号:

TP183

基金项目:

国家自然科学基金资助项目(No. 62102394);福建省科技计划项目(2023N3010)


Graph Structure Motion Segmentation Method for Geometric Information Learning
Author:
Affiliation:

Fund Project:

  • 摘要
  • |
  • 图/表
  • |
  • 访问统计
  • |
  • 参考文献
  • |
  • 相似文献
  • |
  • 引证文献
  • |
  • 资源附件
  • |
  • 文章评论
    摘要:

    针对现有运动分割方法在交通场景下实用性方面的不足,性能和验证时间难以平衡的问题,提出用于几何信息学习的图结构运动分割方法(GS-Net)。GS-Net由点嵌入模块、局部上下文融合模块、全局双边正则化模块和分类模块组成。其中,点嵌入模块将原始关键特征点数据从低维线性难可分的空间映射到高维线性易可分的空间,有利于网络学习图像中运动对象之间的关系;局部上下文融合模块利用双分支图结构分别在特征空间和几何空间提取局部信息,随后将两种类型的信息融合得到更强大的局部特征表征;全局双边正则化模块则利用逐点和逐通道的全局感知来增强局部上下文融合模块得到的局部特征表征;分类模块将前面得到的增强局部特征表征映射回低维分类空间进行分割。GS-Net在KT3DMoSeg数据集的误分类率均值和中值分别为2.47%和0.49%,较于SubspaceNet分别降低8.15%和7.95%;较于SUBSET分别降低7.2%和0.57%。同时,GS-Net在网络推理速度相比SubspaceNet和SUBSET均提升两个数量级;GS-Net在FBMS数据集召回率和F-measure分别为82.53%和81.93%,较于SubspaceNet分别提升13.33%和5.36%,较于SUBSET分别提升9.66%和3.71%。实验结果表明GS-Net能够快速、精确地分割出真实交通场景中的运动物体。

    Abstract:

    The graph-structured motion segmentation method (GS-Net) for geometric information learning is proposed to address the shortcomings of existing motion segmentation methods in terms of their practicality in traffic scenarios, and the difficulty in balancing performance and validation time. GS-Net consists of a point embedding module, a local context fusion module, a global bilateral regularization module, and a classification module. The point embedding module maps the original key feature point data from a low-dimensional linearly difficult-to-differentiate space to a high-dimensional linearly easy-to-differentiate space, which is conducive to the network learning the relationship between moving objects in the image; the local context fusion module utilizes the dual-branching graph structure to extract local information from both the feature space and the geometric space, and then fuses the two types of information to obtain a more powerful local feature representation, The global bilateral regularization module uses point-by-point and channel-by-channel global sensing to enhance the local feature representations obtained by the local context fusion module; the classification module maps the enhanced local feature representations back to the low-dimensional classification space for segmentation. GS-Net"s mean and median misclassification rates on the KT3DMoSeg dataset are 2.47% and 0.49%, respectively, which are 8.15% and 7.95% lower than those of SubspaceNet, and 7.2% and 0.57% lower than those of SUBSET. Meanwhile, GS-Net improves the network inference speed by two orders of magnitude compared to both SubspaceNet and SUBSET. GS-Net"s recall and F-measure on the FBMS dataset are 82.53% and 81.93%, respectively, showing improvements of 13.33% and 5.36% compared to SubspaceNet, and 9.66% and 3.71% compared to SUBSET, respectively. The experimental results demonstrate that GS-Net can quickly and accurately segment moving objects in real traffic scenes.

    参考文献
    相似文献
    引证文献
引用本文
分享
文章指标
  • 点击次数:
  • 下载次数:
  • HTML阅读次数:
  • 引用次数:
历史
  • 收稿日期:2024-06-26
  • 最后修改日期:2024-12-05
  • 录用日期:2024-12-09
  • 在线发布日期:
  • 出版日期:
文章二维码
×
《电子测量与仪器学报》
财务封账不开票通知