首页 | 官方网站   微博 | 高级检索  
     

基于多任务学习的高分辨率遥感影像建筑提取
引用本文:朱盼盼,李帅朋,张立强,李洋.基于多任务学习的高分辨率遥感影像建筑提取[J].地球信息科学,2021,23(3):514-523.
作者姓名:朱盼盼  李帅朋  张立强  李洋
作者单位:1.北京师范大学地理科学学部, 北京 1008752.北京师范大学环境遥感与数字城市北京市重点实验室, 北京 100875
基金项目:国家自然科学基金项目(41371324)。
摘    要:建筑物的自动提取对城市发展与规划、防灾预警等意义重大。当前的建筑物提取研究取得了很好的成果,但现有研究多把建筑提取当成语义分割问题来处理,不能区分不同的建筑个体,且在提取精度方面仍然存在提升的空间。近年来,基于多任务学习的深度学习方法已在计算机视觉领域得到广泛应用,但其在高分辨率遥感影像自动解译任务上的应用还有待进一步发展。本研究借鉴经典的实例分割算法Mask R-CNN和语义分割算法U-Net的思想,设计了一种将语义分割模块植入实例分割框架的深度神经网络结构,利用多种任务之间的信息互补性来提升模型的泛化性能。自底向上的路径增强结构缩短了低层细节信息向上传递的路径。自适应的特征池化使得实例分割网络可以充分利用多尺度信息。在多任务训练模式下完成了对遥感影像中建筑物的自动分割,并在经典的遥感影像数据集SpaceNet上对该方法进行验证。结果表明,本文提出的基于多任务学习的建筑提取方法在巴黎数据集上建筑实例分割精度达到58.8%,在喀土穆数据集上建筑实例分割精度达到60.7%,相比Mask R-CNN和U-Net提升1%~2%。

关 键 词:深度学习  多任务学习  语义分割  实例分割  遥感影像  建筑提取  Mask  R-CNN  U-Net  
收稿时间:2019-12-26

Multitask Learning-based Building Extraction from High-Resolution Remote Sensing Images
ZHU Panpan,LI Shuaipeng,ZHANG Liqiang,LI Yang.Multitask Learning-based Building Extraction from High-Resolution Remote Sensing Images[J].Geo-information Science,2021,23(3):514-523.
Authors:ZHU Panpan  LI Shuaipeng  ZHANG Liqiang  LI Yang
Affiliation:1. Faculty of Geographical Science, Beijing Normal University, Beijing 100875, China2. Beijing Key Laboratory of Environmental Remote Sensing and Digital City, Beijing Normal University, Beijing 100875, China
Abstract:Automatic extraction of buildings is of great significance to urban development and planning, and disaster prevention and early warning. Current researches on building extraction have achieved good results, but the existing research methods often take building extraction as a semantic segmentation problem and cannot distinguish different building individuals. Thus, there is still room of improvement in extraction accuracy. In recent years, deep learning methods based on multitask learning have been widely used in the field of computer vision, but its application in automatic interpretation of high-resolution remote sensing images has not yet further developed. The instance segmentation branch of Mask R-CNN is built on the basis of target detection, and can predict segmentation masks on each region of interest. However, some spatial details and the contextual information of the edge pixels of the region of interest will be lost inevitably. The semantic segmentation task can introduce more contextual information to the network. Therefore, the integration of semantic segmentation and instance segmentation tasks can improve the generalization performance of the whole network. Based on the classic instance segmentation method(Mask R-CNN) and a typical semantic segmentation method(U-Net), this research designs a deep neural network structure which embeds the semantic segmentation module into the instance segmentation framework, and improves the generalization performance of the model by using the information complementarity between various tasks. The bottom-up path augmentation structure shortens the path of lower layers’ information to pass up. The adaptive feature pooling makes it possible for instance segmentation network to make full use of multi-scale information. The automatic building segmentation of remote sensing images is performed in the multi-task training mode and the proposed method is verified on the classic remote sensing image data set(SpaceNet). The result shows that the building instance segmentation accuracy of our proposed method is 58.8% in the Paris data set and 60.7% in the Khartoum data set, increased by1%~2% compared to individual Mask R-CNN and U-Net. The disadvantages of the proposed method are shown in two aspects, one is that the false extraction and missing extraction of small buildings are relatively high, and the other is that the accuracy of building boundary extraction needs to be improved.
Keywords:deep learning  multi-task learning  semantic segmentation  instance segmentation  remote sensing  building extraction  Mask R-CNN  U-Net
本文献已被 CNKI 维普 万方数据 等数据库收录!
点击此处可从《地球信息科学》浏览原始摘要信息
点击此处可从《地球信息科学》下载全文
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司    京ICP备09084417号-23

京公网安备 11010802026262号