Next Article in Journal
Parabolic-Index Ring-Core Fiber Supporting High-Purity Orbital Angular Momentum Modes
Previous Article in Journal
Vehicle Instance Segmentation Polygonal Dataset for a Private Surveillance System
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Query-Based Network for Rural Homestead Extraction from VHR Remote Sensing Images

1
Institute of Agricultural Information, Chinese Academy of Agricultural Sciences, Beijing 100876, China
2
Key Laboratory of Agricultural Blockchain Application, Ministry of Agriculture and Rural Affairs, Beijing 100125, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(7), 3643; https://doi.org/10.3390/s23073643
Submission received: 9 November 2022 / Revised: 15 March 2023 / Accepted: 23 March 2023 / Published: 31 March 2023
(This article belongs to the Section Remote Sensors)

Abstract

:
It is very significant for rural planning to accurately count the number and area of rural homesteads by means of automation. The development of deep learning makes it possible to achieve this goal. At present, many effective works have been conducted to extract building objects from VHR images using semantic segmentation technology, but they do not extract instance objects and do not work for densely distributed and overlapping rural homesteads. Most of the existing mainstream instance segmentation frameworks are based on the top-down structure. The model is complex and requires a large number of manually set thresholds. In order to solve the above difficult problems, we designed a simple query-based instance segmentation framework, QueryFormer, which includes an encoder and a decoder. A multi-scale deformable attention mechanism is incorporated into the encoder, resulting in significant computational savings, while also achieving effective results. In the decoder, we designed multiple groups, and used a Many-to-One label assignment method to make the image feature region be queried faster. Experiments show that our method achieves better performance (52.8AP) than the other most advanced models (+0.8AP) in the task of extracting rural homesteads in dense regions. This study shows that query-based instance segmentation framework has strong application potential in remote sensing images.

1. Introduction

As an important part of China’s rural areas, rural homesteads are directly related to the life and economic development of farmers. In recent years, the Chinese government has begun to gradually attach importance to the management of rural homesteads, and comprehensively grasp the relevant data of rural homesteads through an information survey. With the rapid development of satellites, unmanned aerial vehicles, and other aerospace platforms, a wide range of very high-resolution (VHR) remote sensing images can be obtained. At present, VHR images have been used in a large number of fields, such as cadastral mapping, land classification, road extraction, and rural planning and change detection [1,2,3,4]. VHR images can help us analyze the target objects clearly and accurately. Combined with the advantages of VHR imaging, the current survey method of information such as the quantity and area of homestead has changed from traditional field mapping to manual remote sensing image interpretation, and the rural homestead map spots are extracted by visual interpretation. Although high-resolution remote sensing images can provide clear and large-scale ground information, the distribution characteristics of rural homesteads are dense and large in number. The cost of manually extracting homestead map spots from remote sensing images is high and the efficiency is low, thereby creating difficulties to meet the national demand for large-scale homestead information statistics. Therefore, a fast and accurate automation method is needed to solve this problem.
A remote sensing image has the characteristics of multi-scale, diversity, light sensitivity, complex features, and so on [5]. Making full use of remote sensing image features is the key factor to improve the performance of recognition methods. Some traditional methods achieve the target extraction task by manually extracting morphological indicators, spectral features, and some semantic features of remote sensing images [6,7,8]. However, these indicators are sometimes unreliable. Changes in the shooting angle, illumination conditions, intra-class spectral changes, and inter-class spectral similarity of remote sensing images often limit the application of these methods [9]. AlexNet won the championship in the ImageNet competition in 2012 [10]. After that, deep learning began to show its great potential in the field of images and gradually entered the vision of researchers. People try to apply depth learning technology to image object detection and segmentation, and have achieved great success [11,12]. Compared with traditional algorithms, deep learning does not need to rely on manually extracted prior knowledge and can automatically learn multi-level and multi-dimensional implicit information, which provides an efficient and stable method for interpreting remote sensing images with rich surface feature information. Before 2020, the field of computer vision mainly used CNN as the main framework of deep learning. Here, by stacking convolutional blocks with convolutional kernels of different sizes, the receptive field is gradually increased to extract low-level to high-level semantic information. Many excellent instance segmentation algorithms based on CNN architecture have been proposed. These algorithms are often divided into two methods: top-down and bottom-up. The top-down method is generally to perform target detection first and then segmentation, whereas the bottom-up method is generally to perform semantic segmentation first, and then distinguish different instances through clustering or other measurement methods. However, the post-processing steps of the bottom-up method are often cumbersome, and the accuracy is not high. All the current mainstream methods are top-down methods. The top-down method is divided into the one-stage method and two-stage method according to whether the target detection network needs a proposal region. The two-stage method requires that the candidate boxes are classified, regressed, and masked. Such methods are represented by Mask-RCNN, Cascade-RCNN, HTC, etc. [13,14,15,16]. They can often achieve better accuracy, but the computational cost is large, and the speed is slow. The one-stage method does not need to generate a proposal region, but directly predicts the category, location, and mask of the target. This type of method represents YOLACT, BlendMask, SOLOV1, SOLOV2, etc. [17,18,19,20]. The advantages of this type of method are high efficiency, low computational overhead, but the accuracy cannot be guaranteed. These two types of methods have a common problem. They depend on NMS-related operations, which makes them unable to become end-to-end network architectures. Vision Transformer (VIT), proposed in 2020, is the first model to successfully apply transformer architecture to the field of computer vision [21]. Subsequently, a transformer-based object detection algorithm Detection Transformer (DETR) was proposed [22]. It is the first end-to-end network in the true sense, and it does not need to manually set a threshold to filter out reasonable target boxes and masks. DETR uses a new idea to detect objects, and searches for instance regions in images by querying. This architecture is simpler and more elegant. At the same time, the emergence of DETR also opens a new idea for instance segmentation. Instance segmentation does not need the traditional top-down method of detection and segmentation but can also be completed by finding the mask of the instance region on the image, which is called the query-based method. QueryInst, MaskFormer, Mask2Former, and other advanced instance segmentation are also designed based on the query-based method [23,24,25,26]. This kind of method has high accuracy and is simpler in structure than the traditional top-down instance segmentation method. Moreover, Transformer’s attention mechanism has the ability to capture global context information, so it is more suitable for remote sensing images that have rich context semantic information and complex scenes.
Although the current instance segmentation method has achieved excellent performance, it is rarely used in remote sensing images. Fang et al. designed an attention module and a boundary optimization module to improve the detection performance of small target buildings [27]. Li et al. used the Mask-RCNN framework to identify new and old buildings in rural areas [28]. Wu et al. proposed an Anchor free framework to solve the problem of building instance segmentation [29]. Liu et al. uses an improved HTC algorithm to solve the problem of building instance segmentation [30]. Although these CNN-based methods have good performance, the overall structure is relatively complex, and end-to-end building instance segmentation cannot be achieved. A large number of manually set thresholds, such as anchor’s aspect ratio, confidence, and NMS threshold are required. When buildings show significant shape differences and scale variability, some anchor-free detectors cannot accurately regress the position caused by fixed grid calculation in convolution [9]. Most remote sensing image interpretation tasks, especially building extraction tasks, are based on semantic segmentation models, and most remote sensing image building extraction dataset labels are also designed manually for semantic segmentation [31,32,33,34]. Although the performance of the semantic segmentation model is already well, it can not achieve satisfactory results when encountering very dense and continuous regions in rural areas. As shown in Figure 1, the distribution of buildings in rural areas in China tends to show an aggregation pattern, scattered in different locations in the form of villages, and the distribution of rural homesteads in villages is very dense. It is easy to mistakenly identify two closely connected rural homesteads as one, which will lead to errors in subsequent statistical information. Although there are rare related works using semantic segmentation models to separate adjacent rural homesteads [35], when there is overlap in the space of adjacent rural homesteads, the semantic labels only mark them as the category of homesteads. However, it is impossible to distinguish instance individuals, which has caused huge interference to the statistical quantity and area.
To solve the above problems, we propose a network for rural homestead extraction in very dense areas based on the simpler query-based architecture. The main contributions of this study are summarized as follows:
We propose an end-to-end instance segmentation method based on Transformer architecture and build a multi-scale deformable attention module in encoder to aggregate cross hierarchical global context information to enhance distance dependency, in addition, this greatly reduces the number of parameters and improved the convergence speed;
Group queries are built in the decoder, which changes the One-to-One label assignment method of the original DETR to the Many-to-One label assignment method, greatly improving the convergence speed of the decoder. Moreover, no additional parameters are required in the inference stage;
We propose a method to solve the problem of it being difficult to accurately extract rural homestead due to its dense distribution. The extracted pattern can be used to count the number, area, and other information regarding homesteads.

2. Materials and Methods

2.1. Data Collecting

Our data were collected from UAV images with a spatial resolution of 0.2 m, covering the whole area of Deqing County, Zhejiang Province, China. We used the LabelMe annotation tool to manually mark rural homesteads in a partial cropping area and generate binary mask label images. To generate instance-level masks, we also saved the JSON files generated by the LabelMe tool.

2.2. Methods

2.2.1. Overview of Network Architecture

As shown in Figure 2, the QueryFormer network proposed in this paper is based on Transformer architecture. Aggregation of context information is important for remote sensing images with multi-scale information. The Transformer architecture is spatially capable of long-distance modeling, captures context-dependent, long-distance relationships, and is well suited for remote sensing imagery. In QueryFormer, there is an encoder and a decoder. To extract features, we utilize the Swin-Transformer network as the backbone, which has recently demonstrated excellent performance [36]. The Swin-Transformer network addresses the limitations of the original Vision Transformer, which struggles to extract multiscale information and is not well-suited for small targets in images. After extracting the information using Swin-Transformer, the feature maps with four resolutions are obtained, which are 1/4, 1/8, 1/16, and 1/32. High-resolution feature maps can extract underlying information such as image boundaries and textures, and are friendly to small targets, whereas low-resolution images have strong semantic features. Feature maps with 1/8, 1/16, and 1/32 down-sampling resolution were input into the Multi-Scale Deformable Attention module (see Section 2.2.2) to obtain a self-attention feature map. In order to obtain high-quality mask feature maps, the self-attention feature maps of three resolutions were up sampled to the highest resolution feature map and added together to obtain high-resolution feature maps with rich boundary and semantic information. QueryFormer decoder is similar to DETR decoder, and we first initialize group queries embedding (see Section 2.2.4), which is shaped as K × N × C, where K is the number of groups, N is the number of queries, and C is the number of embedding dimensions. In this experiment, we set K to 5 and N to 300. In the decoder, the standard self-attention module and cross-attention module (see Section 2.2.3) are used to compute from multi-scale image features and N learnable positional embeddings from its output. Each mask query predicted by QueryFormer is encoded using N per-mask embeddings Q R C Q × N of dimension CQ, which capture global information. The query results are then passed through a linear classifier layer (FFN layer) to generate decoder embeddings. These decoder embeddings are further processed by an MLP layer and a SoftMax activation layer to produce class probability embeddings and mask embeddings, respectively (as described in Section 2.2.4).

2.2.2. Multi-Scale Deformable Attention Module

In order to enhance the effect of feature extraction and aggregate context information, we add an attention mechanism behind the backbone. At present, the mainstream methods to solve the task of target detection and instance segmentation are a process from dense to sparse. They all need dense prior knowledge (anchor, reference points, etc.), but there are some problems with dense prior. Firstly, many similar results will be detected, and post-processing methods such as threshold and NMS are required to filter. Secondly, the detection results are closely related to a priori (the number and size of anchors, the degree of confidentiality of reference points, the number of proposals generated, etc.). These super parameters need to be carefully initialized. Carion et al. proposed a novel object detection model, called DETR, which employs an encoder–decoder structure combined with the powerful global modeling capability of the Transformer architecture [22]. By eliminating the need for manual design of anchors, reference points, and other prior knowledge, as well as post-processing techniques such as NMS, they were able to create the first completely end-to-end object detector. Although the DETR has better performance and simpler ideas, there are still some problems. Firstly, compared with the mainstream object detectors, the DETR needs more time to train to converge. Secondly, the effect of DETR on small target detection is not effective. The main reasons for the above problems are as follows. Firstly, the Transformer structure has defects in processing image feature maps. During initialization, the attention module applies nearly consistent attention weights to all pixels in the feature map. A long training time is necessary to focus on a sparse and meaningful position. On the other hand, the calculation of the weight of attention in the Transformer encoder is a secondary calculation of the number of pixels. Therefore, processing high-resolution feature maps has very high computational and memory complexity. Secondly, target detectors with good performance use multi-scale features to detect small objects from high-resolution feature maps, which make DETR highly complex and unacceptable. In order to solve the above problems, encouraged by Deformable DETR [37], we designed a multi-scale Deformable attention (MSDeformattn) module in the network. The MSDeformattn module combines Deformable Evolution’s sparse feature space sampling and Transformer’s global modeling capabilities to alleviate the slow convergence and high complexity of the DETR algorithm.
Give a scale feature map x R C × H × W , the expression of deformable attention is:
D e f o r m A t t n z q , p q , x = m = 1 M W m k = 1 K A m q k · W m x p q + Δ p m q k
where q is the index of a feature location on the feature map x, zq is the corresponding feature vector on x for the index q, and pq is the corresponding two-dimensional reference point on x for the index q. M represents the number of attention heads, K represents the number of sampled keys, W and W represent two different linear projection layers. Amqk denotes attention weights at the kth sampling point of the mth attention head corresponding to index q, pmqk denotes learnable sampling offsets at the kth sampling point of the mth attention head corresponding to index q, and x p q + Δ p m q k is calculated by bilinear interpolation. Figure 3 shows that there are two attention heads (M = 2) and three sampling offset points (K = 3) on a single scale. First, flatten the single scale feature map x R C × H × W shape to C × N , where N = H × W and C is the number of original channels. Unlike the self-attention module in the Vision Transformer, the query feature is not generated by linear projection of the original sequence, but the original sequence with position encoding. Then the query feature is passed through a linear projection layer with an output channel of 2MK to obtain the encoded sampling offsets Δ p m q k . For generating the encoded attention weights A m q k 0 , 1 , the query feature is passed through another linear projection layer with output channel of MK and SoftMax activation layer. Multiply A m q k with the characteristic value of a resampled offset point through the linear projection layer to obtain the feature map after deformable attention.
The multi-scale deformable attention module is to sample LK points on the feature map of multiple levels, then perform the above operations separately. Finally, the feature maps of all scales are added together to get the multiscale feature map, which is expressed as
M S D e f o r m A t t n z q , p q , x l l = 1 L = m = 1 M W m l = 1 L k = 1 K A m l q k · W m x l ϕ l p q + Δ p m l q k
where L denotes L the number of levels, and other parameters is the same as the parameters in Expression (2).

2.2.3. Self-Attention Module and Cross-Attention Module

Inspired by DETR structure, we add two attention modules to decoder, which are self-attention module and cross-attention module. Its structure is shown in Figure 4. Their expressions are shown in Equations (3) and (4).
S e l f - A t t e n t i o n Q s , K s , V s = s o f t max Q s K s T d k V s
C r o s s - A t t e n t i o n Q c , K c , V c = s o f t max Q c K c T d k V c
First, query embeds in each group initialized to N × C are divided into three same parts: query, key, and value. Their shapes are also N × C. Query and key are added with the same query position embeds, and then they are separately passed through a linear projection layer to obtain Qs, Ks. Enter value into a linear projection layer to get Vs. K is transposed to a sequence with a shape of C × N , multiplied by Q, attention weights are obtained through the SoftMax activation function, and V and attention weights are multiplied to get an output sequence with a shape of N × N . By adding query embeds and outputs and passing through the LayerNorm layer, the weighted query feature is obtained, and the self-attention operation is completed. Following this is the cross-attention module, which adds query position embedding to the query feature obtained from the self-attention module and obtains Q c K N × C through a linear projection layer, whereas the key and value in the cross-attention come from the three-layer feature map in the encoder. K c H W × C is obtained by adding key position embedding and passing through a linear projection layer. V C H W × C is formed by value passing directly through a linear projection layer. The steps to obtain attention weights and features follow the same pattern as self-attention. The result’s shape is KN × C after cross-attention module.

2.2.4. Group Queries Assignment

In the traditional CNN-based object detection algorithm, a feature region is usually mapped by multiple proposal boxes, and then the prediction box closest to the label is obtained through NMS or other post-processing methods. This is equivalent to a target instance being predicted by multiple anchors, which can focus on the target region more quickly and is a Many-to-One matching method. The end-to-end method DETR discards the NMS operation, but it cannot perform Many-to-One label matching, which makes the convergence speed very slow. Therefore, we use multiple groups of queries to perform many to one matching. This matching method only occurs in the training process, and in the inference stage, it is only necessary to take the first group of queries for forward propagation, so that the calculation consumption during reasoning is not increased. As shown in Figure 5, there are K groups of queries. In the forward propagation process, K groups of predictions are obtained. Each group of predictions consists of category probability prediction vectors and mask embeds. The mask embeds shape is N × C, it multiplies the high-resolution feature map with the shape of C × 4/H × 4/W in encoder to obtain the shape of N × 4/H × 4/W mask prediction. Each group of prediction matches the label in the One-to-One way, and then updates the weight through the loss function. In this way, a region in the label is mapped by the feature regions extracted by K queries, so as to achieve the goal of many to one matching.

2.2.5. Loss Design

As a network architecture for end-to-end instance segmentation, QueryFormer’s loss function consists of two parts: classification loss and mask loss. The classification loss uses focal loss, whereas the mask loss uses the weighted sum of cross entropy loss and dice loss [38].
For remote sensing images, there is often a large number of unbalanced categories, so focal loss is used to deal with the unbalance distribution of categories. The calculation formula of focal loss is shown in Equation (5).
F L p t = t 1 p t γ log p t
where, t 0 , 1 is a super parameter used to control the imbalance of positive and negative samples, γ 0 , 5 is a super parameter to balance difficult samples and simple samples, and p t is the probability distribution of prediction of different categories. For positive samples, when P is large, the category confidence of prediction is high, and it is a simple sample, so 1 p t γ is small, and the loss value is small. On the contrary, when p is small, the sample confidence is low, and the loss value is large. For negative samples, when P is small, the prediction category is more inclined to negative samples, and the loss value is small. On the contrary, when P is large, the loss value is large. By adjusting γ to reduce the impact of simple samples, the network is made pay more attention to difficult samples. QueryFormer outputs a mask shape of B × Q × H × W at the last layer of decoder in one group (noted that we only introduce one group loss here, and is same as other group), where B is batch size, Q is the number of queries, and H and W are the height and width of a quarter of the input image. For most segmentation networks, the output shape is B × C × H × W , C is the classes numbers, and the quantity of Q is much larger than the quantity of C. Using the mask result which shape is B × Q × H × W to calculate the loss will result in enormous training memory consumption. Inspired by some efficient work, we calculate the loss function by selecting K sample points for each H × W feature map based on the uncertainty of the current pixel instead of the entire feature map [16,25], where the uncertainty is measured based on the pixel prediction confidence. The formula for calculating classification loss is shown in Equation (5).

3. Results

3.1. Dataset

The original image datasets were labeled by LabelMe to general semantic masks, JSON file was used to assign an instance to each mask object, and then we generated the instance masks. Then, we produced the COCO format annotation JSON files. The semantic masks generated from LabelMe tool, and the instance masks generated by converting are shown in Figure 5 below. The size of each image data used for training and testing is more than 10,000 × 10,000, which cannot be put into the GPU devices for calculation. Therefore, we use the overlapping clipping method to cut out several 640 ∗ 640 images, with the size of the middle overlapping part of 200. If the length remaining after cropping the last image of a row or column is less than 640, the final image of the row or column will be cropped from its size minus 640 to its original size. Finally, the number of cut sample data used for training is 22,338, and the number of data used for testing is 3630. In the training stage, we use data augmentation methods such as random flipping, random scaling and clipping, and random color change for cropped images.

3.2. Experimental Environment and Details

The experimental environment is based on the Ubuntu 18.04 system, using Python 3.8 and PyTorch 1.12.1 to build models and train. The hardware conditions are Intel (R) Xeon (R) CPU E5-2678 v3 @ 2.50G Hz, 256G RAM, and 8 × 24G RTX3090 GPUs. At the beginning of training, we set the initial learning rate to 0.0001, and use warmup strategy to increase the learning rate [39]. After 10,000 steps, we used the poly method to decrease learning rate [40]. The poly method is shown in Equation (6).
l r = b a s e _ l r × 1 c u r _ i t e r s max _ i t e r s p o w e r
where base_lr is the last learning rate, and cur_iters is the current number of steps, max_iters is the total number of training steps, and the value of power is set to 0.9. The optimizer we chose to update the model’s weights is AdamW [41]. We set the batch size per GPU is 2, and the total batch size is 16.

3.3. Comparison of Experimental Results

In order to prove the effectiveness of our proposed network, we compared the results with those of other SOTA models under the same experimental conditions, including Mask-RCNN, Cascade-RCNN, QueryInst, and Mask2Former. The first three are instance segmentation models based on CNN architecture, and they follow the top-down detection paradigm. Moreso, the last two are models based on the Transformer architecture and follow the query method to obtain the prospect target. All methods adopt ResNet [42] and Swin-Transformer as the backbone, and we selected ResNet101 and Swin-B for comparison. To make a fair comparison, we list the methods in which different algorithms query instances. The CNN-based method uses anchor boxes, whereas the Transformer-based method uses 300 randomly initialized learnable queries (QueryFormer uses five groups for training and one group for inferencing). Table 1 shows the performance comparison results of QueryFormer with other SOTA methods. In contrast experiments, we only evaluate the quality of the instance mask. The integrity of the Transformer-based method is better than that of the CNN-based method. In the experiment with Swin-B as the backbone, the average mask precision of our QueryFormer was 52.8%, which is 0.8% higher than Mask2Former and 1.7% higher than HTC. The average precision of medium-sized targets (APM) was 1.0%, higher than Mask2Former and 2.1% higher than HTC. The average precision of large targets (APL) was 0.9% higher than Mask2Former and 2.4% higher than HTC. In the comparison of indicators APS, which are used to evaluate the segmentation quantitative of small targets, the highest one was HTC, and QueryFormer was 0.2% lower than QueryInst.
In order to better compare the effect of our Transformer-based architecture model with the traditional CNN segmentation model, we selected the best performing models in two architectures, QueryFormer and HTC. Figure 6 shows the comparison between QueryFormer and HTC in complex scenes. The left column is the original image, the middle column is the HTC segmentation effect, and the right column is the QueryFormer segmentation effect. In the first and second scenes, the red and white boxes represent the recognition effect of the two models on the shadow parts. In the first scene, the black shadow in the red box is actually a courtyard, whereas the HTC model recognizes the black shadow as a homestead, and QueryFormer does not make such an error. In the second scene, the shadow in the red box includes the homestead sheltered by vegetation. Although the HTC model recognizes the homestead in the shadow, it cannot correctly represent the mask. QueryFormer accurately represents the mask of the house site that is not covered by vegetation in the shadow. Some edges of a homestead in the white box are in shadow. The HTC model fails to correctly identify the edges of the homestead contained in the shadow, resulting in a missing mask, but QueryFormer avoids this problem.

3.4. Performance Effect and Transferability of QueryFormer

Figure 7 shows the effect of Swin-Transformer-backbone-based QueryFormer in a rural homestead extraction task. Whether in dense or sparse areas, QueryFormer can clearly and correctly extract the homestead. In more complex scenes, QueryFormer can precisely query each instance targets, identify objects of different sizes almost without errors, and accurately separate the background region.
We found an interesting phenomenon in the experiment. QueryFormer also performs well on other untrained datasets. As shown in Figure 8, the first image is from a rural area, but the resolution is different from the data we trained, and the overall image is brighter due to lighting. However, QueryFormer better identified each instance of a building, but some obvious errors occurred in HTC which perform best in the CNN structure, failing to identify the buildings in the red box and yellow box. The resolution of the second image is also different from the training data, and there are some differences in roof shape and pixel distribution. QueryFormer is good at identifying most targets, but HTC misses identifying many buildings, such as foreground targets in red and yellow boxes. The third and fourth images are buildings in urban areas. There are obvious differences between building styles and buildings in rural areas, and there are many unknown distributions in the background pixels. It is difficult to transfer from rural areas to urban area. For example, in the fourth image, QueryFormer misidentifies a road as a target in a light green box, but HTC does not. However, in other areas, such as red and yellow boxes, HTC has obvious recognition errors, and QueryFormer can correctly identify most of the targets. The comparison results show the advantages of our Transformer-based architecture in transferability. The foreground features are highlighted by establishing local and global context relations. When the building texture, shape, and brightness distribution are different from the training data, the object type can also be inferred through the context relations, and then each target region can be obtained by querying.

3.5. Hyperparameters Contrast Experiment

In the QueryFormer model structure, some hyperparameters settings play a key role, which directly determines the effect of the model. In order to get the best hyperparameters, we set up several ablation experiments to verify the most appropriate choice of hyperparameters. In the ablation experiment, we used Swin-Transformer as the backbone and trained 12 epochs without special instructions.
The effectiveness of the query quantity. In order to compare the influence of different query numbers in the Transformer decoder on the results, we selected four query numbers for the comparative experiment (100–400). Please note that the queries number here are the quantity of queries in each group; we use five groups for training and one group for inferencing. As shown in Table 2, the more queries there are, the better effect is, but the calculating costs also increase substantially. For example, when we calculate the self-attention, the product shape of q and k is N × N , and N is the number of queries. When N increases from 300 to 400, N × N increases by 10,000, which brings a lot of computing overheads. In addition, after the number of queries reaches to 300, the effect does not increase significantly. When the number of queries is 400, each evaluation index has only increased by 0.3, 0.2, 0, and 0.4. Considering the effect and calculation cost, we chose 300 queries for the experiment.
The effectiveness of the group quantity. To compare the impact of the number of grouped queries on the results, we set five groups number for comparative experiment (1, 2, 3, 4, 5). When the group number equals 1, it is the same as the original DETR. We only use the first set of queries in the inference process. As shown in Table 3, the more groups there are, the better effect is. The model trained with group query obtains 2.7 mAP and 2.0 mAP gains over the baseline under the 12 epochs and 50 epochs. However, in fact, the calculating costs are also increasing a lot in training stage. When the groups number increases from 1 to 5, the total queries number increases from 300 to 1500. Although the inference stage does not increase the calculation cost, it will bring additional costs in the training stage. Considering that increasing the number of groups does not bring a significant rise point, the number of groups selected in our experiment is five. Moreover, during training for 50 epochs, the mean average precision (mAP) is observed to be 2% lower when the number of groups is 1 (i.e., the original DETR decoder) as compared to when the number of groups is 5. This suggests that grouping queries can enhance accuracy and expedite model training.

4. Discussion

4.1. What Deformable Attention Learns in QueryFormer Encoder

Deformable attention module saves a lot of network training consumption and helps the network focus on the interest region faster. We visualize what each sample point of each attention head learns when the sample point is 4 in Figure 9 and Figure 10. Figure 9 and Figure 10 are heat maps of the attention distribution of highest resolution (8×) and lowest resolution (32×), respectively. The highlighted part of the color shows that the higher the attention weight, the more attention is paid to this region. The attention distribution deviation at high resolution is significant, and the basic contour of the target can be seen through the attention distribution, whereas the attention features extracted at low resolution are more abstract. In order to better visualize the results, we superimpose the attention map at low resolution and the original image.
In Figure 9, we observed that most of the points pay more attention to the details of the image and have higher weight on the target contour of the homestead object, such as the offset points corresponding to the highlighted parts of the homestead contour in head2, head5, head6, head7, and head8. We find that head8 also pays attention to the contour information in all directions of the homestead. Some points focus on the background information of the image, such as point1 in head1, head2, head6, and head7. In Figure 10, most of the points pay more attention to the advanced information of the image and have higher weight on the semantic part of the homestead, such as the points in head2, head4, head5, head7, and head8, which highlight the homestead object part. At the same time, some points focus on the background information of the image, such as point1 in head1, and head2. It is worth noting that the distribution of the attention weights of point2 and point3 of head3 in the attention heat map of the lowest resolution is somewhat special. We think it is a response to the context relationship between some homestead pixels and background pixels.

4.2. What Query Learns in QueryFormer Decoder

Similar to the region proposal proposed by Faster-RCNN, learnable query in the cross-attention module of the Transformer decoder learns the masks of the interest region proposals from the feature map in the encoder and gives higher weight to the target reg ion. To explore what features query has learned in the encoder, we selected the first 15 of the 300 queries in the first group for visualization experiments in Figure 11. We observe that the weights of the first 15 queries are distributed in different regions, corresponding to different homestead target objects, and the contour of the highlighted part of the attention is a complete homestead, which shows that each learnable query has completely queried the foreground target information of different regions, which can be used as the region proposal masks.

4.3. What Group Queries Learns in QueryFormer Decoder

Similar to the strategy of anchor-based detection algorithms, such as Faster RCNN in sample assignment, group queries assignment multiple positive objects to the same GT, and multiple parallel queries can quickly find the foreground region of interest. To investigate the learning outcomes of various groups’ queries, we chose the first group’s queries and one from among them. Then, we examined the attention distribution of the corresponding queries in the two groups. As shown in Figure 12, the highlighted part represents the place with high attention weight, and the yellow box is the feature area extracted from the Nth query of the two groups. By comparison, as shown in Figure 13, it is found that the region features learned by the query at the Nth position of the two groups are consistent, indicating that the information of multiple positive sample label allocation matching is consistent. In the reasoning stage, only the query features of the first group are required, and no subsequent NMS operations are required.

5. Conclusions

In this study, we propose a completely end-to-end automatic instance segmentation network QueryFormer based on Transformer architecture, which can provide intelligent decision analysis of rural homesteads. QueryFormer combines the advantages of DETR. Instead of using the current mainstream instance segmentation network idea of first detection and then segmentation, it uses multiple groups of embedded vectors to directly query the characteristics of the current region. The encoder of QueryFormer uses multi-scale deformable attention module instead of the traditional self-attention module, so that the network can learn more representative sampling points, which greatly reduces the computing cost. In order to solve the characteristics of slow and long training times of the DETR-based network, multiple groups of queries are used for parallel computing in the decoder, and the one-to-one positive sample assignment calculation similar to the DETR is performed in each group. In each group, a GT is assigned to a query, and when there are K groups, a GT can be calculated by K queries. Then, K queries are used to learn the characteristics of the region from the feature maps in the encoder, and generates K proposal regions. In the inference process, the calculation will not be increased, and only the first group of queries need to be use for generating predictions. The experiment shows that our QueryFormer achieves the best performance with many state-of-the-art instance segmentation models, and when using multiple groups query to train 50 epochs, the mAP, APS, APM, and APL of our model are 52.8, 29.9, 58.6, and 64.7, respectively. Our method has been verified to have the best results and better transferability among various evaluation indicators, and has been verified to have faster convergence speed and better effect when using group queries. It can be well applied to the extraction of rural homestead in dense regions and be used for subsequent accurate quantitative analysis and visualization.

Author Contributions

R.W. and B.F. conceived and conducted the experiments and performed the data analysis. R.Y. assisted in collating experiment data. R.W. wrote the article. Y.W. and B.F. revised the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Innovation Research Fund of Agricultural Information Institute of CAAS, China (CAAS-ASTIP-2016-AII) and Central Public-interest Scientific Institution Basal Research Fund (No. JBYW-AII-2022-02).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data and code for classification are available at the following GitHub repository: https://github.com/wr19960001/QueryFormer (accessed on 30 December 2022).

Acknowledgments

We thank Deqing County Bureau of Agriculture and Rural Areas, Huzhou City, Zhejiang Province, for the data provided.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Chaib, S.; Liu, H.; Gu, Y.; Yao, H. Deep Feature Fusion for VHR Remote Sensing Scene Classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 4775–4784. [Google Scholar] [CrossRef]
  2. Zhu, Q.; Zhang, Y.; Wang, L.; Zhong, Y.; Guan, Q.; Lu, X.; Zhang, L.; Li, D. A Global Context-aware and Batch-independent Network for road extraction from VHR satellite imagery. ISPRS J. Photogramm. Remote Sens. 2021, 175, 353–365. [Google Scholar] [CrossRef]
  3. Lv, Z.; Liu, T.; Benediktsson, J.A. Object-Oriented Key Point Vector Distance for Binary Land Cover Change Detection Using VHR Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2020, 58, 6524–6533. [Google Scholar] [CrossRef]
  4. Yi, Y.; Zhang, Z.; Zhang, W.; Zhang, C.; Li, W.; Zhao, T. Semantic Segmentation of Urban Buildings from VHR Remote Sensing Imagery Using a Deep Convolutional Neural Network. Remote Sens. 2019, 11, 1774. [Google Scholar] [CrossRef] [Green Version]
  5. Ghassemian, H. A review of remote sensing image fusion methods. Inf. Fusion 2016, 32, 75–89. [Google Scholar] [CrossRef]
  6. Benediktsson, J.A.; Pesaresi, M.; Amason, K. Classification and feature extraction for remote sensing images from urban areas based on morphological transformations. IEEE Trans. Geosci. Remote Sens. 2003, 41, 1940–1949. [Google Scholar] [CrossRef] [Green Version]
  7. Fauvel, M.; Benediktsson, J.A.; Chanussot, J.; Sveinsson, J.R. Spectral and Spatial Classification of Hyperspectral Data Using SVMs and Morphological Profiles. IEEE Trans. Geosci. Remote Sens. 2008, 46, 3804–3814. [Google Scholar] [CrossRef] [Green Version]
  8. Du, S.; Zhang, F.; Zhang, X. Semantic classification of urban buildings combining VHR image and GIS data: An improved random forest approach. ISPRS J. Photogramm. Remote Sens. 2015, 105, 107–119. [Google Scholar] [CrossRef]
  9. Yuan, Q.; Mohd Shafri, H.Z. Multi-Modal Feature Fusion Network with Adaptive Center Point Detector for Building Instance Extraction. Remote Sens. 2022, 14, 4920. [Google Scholar] [CrossRef]
  10. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25. [Google Scholar] [CrossRef] [Green Version]
  11. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [Green Version]
  12. Long, J.; Shelhamer, E.; Darrell, T. Fully Convolutional Networks for Semantic Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 39, 640–651. [Google Scholar]
  13. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. (Eds.) Mask r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017. [Google Scholar]
  14. Cai, Z.; Vasconcelos, N. (Eds.) Cascade r-cnn: Delving into high quality object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  15. Chen, K.; Pang, J.; Wang, J.; Xiong, Y.; Li, X.; Sun, S.; Feng, W.; Liu, Z.; Shi, J.; Ouyang, W.; et al. Hybrid task cascade for instance segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  16. Kirillov, A.; Wu, Y.; He, K.; Girshick, R. PointRend: Image Segmentation as Rendering. In Proceedings of the Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  17. Bolya, D.; Zhou, C.; Xiao, F.; Lee, Y.J. (Eds.) Yolact: Real-time instance segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  18. Chen, H.; Sun, K.; Tian, Z.; Shen, C.; Huang, Y.; Yan, Y. (Eds.) Blendmask: Top-down meets bottom-up for instance segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  19. Wang, X.; Kong, T.; Shen, C.; Jiang, Y.; Li, L. SOLO: Segmenting Objects by Locations. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020. [Google Scholar]
  20. Wang, X.; Zhang, R.; Kong, T.; Li, L.; Shen, C. SOLOv2: Dynamic and Fast Instance Segmentation. Adv. Neural Inf. Process. Syst. 2020, 33, 17721–17732. [Google Scholar]
  21. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image is Worth 16 × 16 Words: Transformers for Image Recognition at Scale. arXiv 2021, arXiv:2010.11929. [Google Scholar]
  22. Kirillov, A.; Usunier, N.; Carion, N.; Zagoruyko, S.; Synnaeve, G.; Massa, F. End-to-End Object Detection with Transformers. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020. [Google Scholar]
  23. Fang, Y.; Yang, S.; Wang, X.; Li, Y.; Fang, C.; Shan, Y.; Feng, B.; Liu, W. Instances as Queries. In Proceedings of the International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021. [Google Scholar]
  24. Cheng, B.; Schwing, A.; Kirillov, A. Per-pixel classification is not all you need for semantic segmentation. Adv. Neural Inf. Process. Syst. 2021, 34, 17864–17875. [Google Scholar]
  25. Cheng, B.; Misra, I.; Schwing, A.G.; Kirillov, A.; Girdhar, R. (Eds.) Masked-attention mask transformer for universal image segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 21–24 June 2022. [Google Scholar]
  26. Dong, B.; Zeng, F.; Wang, T.; Zhang, X.; Wei, Y. Solq: Segmenting objects by learning queries. Adv. Neural Inf. Process. Syst. 2021, 34, 21898–21909. [Google Scholar]
  27. Fang, F.; Wu, K.; Liu, Y.; Li, S.; Wan, B.; Chen, Y.; Zheng, D. A Coarse-to-Fine Contour Optimization Network for Extracting Building Instances from High-Resolution Remote Sensing Imagery. Remote Sens. 2021, 13, 3814. [Google Scholar] [CrossRef]
  28. Li, Y.; Xu, W.; Chen, H.; Jiang, J.; Li, X. A Novel Framework Based on Mask R-CNN and Histogram Thresholding for Scalable Segmentation of New and Old Rural Buildings. Remote Sens. 2021, 13, 1070. [Google Scholar] [CrossRef]
  29. Wu, T.; Hu, Y.; Peng, L.; Chen, R. Improved Anchor-Free Instance Segmentation for Building Extraction from High-Resolution Remote Sensing Images. Remote Sens. 2020, 12, 2910. [Google Scholar] [CrossRef]
  30. Liu, X.; Chen, Y.; Wei, M.; Wang, C.; Gonçalves, W.N.; Marcato, J.; Li, J. Building Instance Extraction Method Based on Improved Hybrid Task Cascade. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  31. Ji, S.; Wei, S.; Lu, M. Fully Convolutional Networks for Multisource Building Extraction From an Open Aerial and Satellite Imagery Data Set. IEEE Trans. Geosci. Remote Sens. 2019, 57, 574–586. [Google Scholar] [CrossRef]
  32. Deng, W.; Shi, Q.; Li, J. Attention-Gate-Based Encoder–Decoder Network for Automatical Building Extraction. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 2611–2620. [Google Scholar] [CrossRef]
  33. Zhou, J.; Liu, Y.; Nie, G.; Cheng, H.; Yang, X.; Chen, X.; Gross, L. Building Extraction and Floor Area Estimation at the Village Level in Rural China Via a Comprehensive Method Integrating UAV Photogrammetry and the Novel EDSANet. Remote Sens. 2022, 14, 5175. [Google Scholar] [CrossRef]
  34. Shi, X.; Huang, H.; Pu, C.; Yang, Y.; Xue, J. CSA-UNet: Channel-Spatial Attention-Based Encoder–Decoder Network for Rural Blue-Roofed Building Extraction From UAV Imagery. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  35. Wei, R.; Fan, B.; Wang, Y.; Zhou, A.; Zhao, Z. MBNet: Multi-Branch Network for Extraction of Rural Homesteads Based on Aerial Images. Remote Sens. 2022, 14, 2443. [Google Scholar] [CrossRef]
  36. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. (Eds.) Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021. [Google Scholar]
  37. Zhu, X.; Su, W.; Lu, L.; Li, B.; Wang, X.; Dai, J. Deformable DETR: Deformable Transformers for End-to-End Object Detection. arXiv 2020, arXiv:2010.04159. [Google Scholar]
  38. Milletari, F.; Navab, N.; Ahmadi, S.-A. V-Net: Fully Convolutional Neural Networks for Volumetric Medical Image Segmentation. In Proceedings of the International Conference on 3d Vision, Stanford, CA, USA, 25–28 October 2016. [Google Scholar]
  39. Xiong, R.; Yang, Y.; He, D.; Zheng, K.; Zheng, S.; Xing, C.; Zhang, H.; Lan, Y.; Wang, L.; Liu, T. On Layer Normalization in the Transformer Architecture. In Proceedings of the International Conference on Machine Learning, Virtual, 13–18 July 2020. [Google Scholar]
  40. Chen, L.-C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 40, 834–848. [Google Scholar] [CrossRef] [Green Version]
  41. Loshchilov, I.; Hutter, F. Decoupled Weight Decay Regularization. arXiv 2018, arXiv:1711.05101. [Google Scholar]
  42. He, K.; Zhang, X.; Ren, S.; Sun, J. (Eds.) Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
Figure 1. Semantic labels and instance labels of rural homestead in dense areas.
Figure 1. Semantic labels and instance labels of rural homestead in dense areas.
Sensors 23 03643 g001
Figure 2. Architecture of QueryFormer.
Figure 2. Architecture of QueryFormer.
Sensors 23 03643 g002
Figure 3. The structure of the single scale deformable attention module.
Figure 3. The structure of the single scale deformable attention module.
Sensors 23 03643 g003
Figure 4. Self-Attention Module In One Group and Cross-Attention Module.
Figure 4. Self-Attention Module In One Group and Cross-Attention Module.
Sensors 23 03643 g004
Figure 5. Group queries assignment in training stage.
Figure 5. Group queries assignment in training stage.
Sensors 23 03643 g005
Figure 6. Examples of training and testing data sets.
Figure 6. Examples of training and testing data sets.
Sensors 23 03643 g006
Figure 7. Visual comparison of the prediction results of QueryFormer and HTC. The first column is the original image, the middle column is the QueryFormer result, and the third column is the HTC result.
Figure 7. Visual comparison of the prediction results of QueryFormer and HTC. The first column is the original image, the middle column is the QueryFormer result, and the third column is the HTC result.
Sensors 23 03643 g007
Figure 8. Visualization of QueryFormer instance segmentation results based on the Swin-Transformer-based backbone.
Figure 8. Visualization of QueryFormer instance segmentation results based on the Swin-Transformer-based backbone.
Sensors 23 03643 g008
Figure 9. QueryFormer and HTC are tested on other untrained data. The first line is the original image, the second line is the QueryFormer result, and the third line is the HTC result.
Figure 9. QueryFormer and HTC are tested on other untrained data. The first line is the original image, the second line is the QueryFormer result, and the third line is the HTC result.
Sensors 23 03643 g009
Figure 10. Deformable attention visual analysis of the highest resolution feature map (8×). Each column represents each head feature, and each row represents each offset sampling point feature.
Figure 10. Deformable attention visual analysis of the highest resolution feature map (8×). Each column represents each head feature, and each row represents each offset sampling point feature.
Sensors 23 03643 g010
Figure 11. Deformable attention visual analysis of the lowest resolution feature map (32×). Each column represents each head feature, and each row represents each offset sampling point feature.
Figure 11. Deformable attention visual analysis of the lowest resolution feature map (32×). Each column represents each head feature, and each row represents each offset sampling point feature.
Sensors 23 03643 g011
Figure 12. Learnable queries visual analysis of the highest resolution feature map (8×). Each column represents each head feature, and each row represents each offset sampling point feature.
Figure 12. Learnable queries visual analysis of the highest resolution feature map (8×). Each column represents each head feature, and each row represents each offset sampling point feature.
Sensors 23 03643 g012
Figure 13. Group queries visual analysis of the highest resolution feature map (8×). Each column represents each head feature, and each row represents each offset sampling point feature.
Figure 13. Group queries visual analysis of the highest resolution feature map (8×). Each column represents each head feature, and each row represents each offset sampling point feature.
Sensors 23 03643 g013
Table 1. Quantitative evaluation (%) of several SOTA algorithms with different backbones on remote sensing images in the test set. Values in bold are the best.
Table 1. Quantitative evaluation (%) of several SOTA algorithms with different backbones on remote sensing images in the test set. Values in bold are the best.
MethodBackboneQuery MethodAPAPSAPMAPL
Mask-RCNNR101Anchors39.823.848.351.1
Cascade-RCNNR101Anchors44.525.951.853.5
HTCR101Anchors47.128.653.359.9
QueryInstR101300 Queries46.528.254.160.2
Mask2FormerR101300 Queries47.727.454.860.8
QueryFormerR101300 Queries48.327.955.761.6
Mask-RCNNSwin-BAnchors44.227.154.159.4
Cascade-RCNNSwin-BAnchors46.727.955.761.4
HTCSwin-BAnchors51.130.456.562.3
QueryInstSwin-B300 Queries50.330.156.962.5
Mask2FormerSwin-B300 Queries52.029.357.663.8
QueryFormerSwin-B300 Queries52.829.958.664.7
Table 2. Comparative experimental results of query quantity.
Table 2. Comparative experimental results of query quantity.
ModelQueriesSchedulesAPAPSAPMAPL
QueryFormer10012e37.616.144.747.3
QueryFormer20012e39.218.947.852.0
QueryFormer30012e41.420.649.553.7
QueryFormer40012e41.720.849.554.1
Table 3. Comparative experimental results of group quantity.
Table 3. Comparative experimental results of group quantity.
ModelGroupsSchedulesAPAPSAPMAPL
QueryFormer112e38.718.446.149.7
QueryFormer212e40.519.447.851.3
QueryFormer312e40.819.948.452.4
QueryFormer412e41.120.349.153.2
QueryFormer512e41.420.649.553.7
QueryFormer150e50.827.155.461.9
QueryFormer550e52.829.958.664.7
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wei, R.; Fan, B.; Wang, Y.; Yang, R. A Query-Based Network for Rural Homestead Extraction from VHR Remote Sensing Images. Sensors 2023, 23, 3643. https://doi.org/10.3390/s23073643

AMA Style

Wei R, Fan B, Wang Y, Yang R. A Query-Based Network for Rural Homestead Extraction from VHR Remote Sensing Images. Sensors. 2023; 23(7):3643. https://doi.org/10.3390/s23073643

Chicago/Turabian Style

Wei, Ren, Beilei Fan, Yuting Wang, and Rongchao Yang. 2023. "A Query-Based Network for Rural Homestead Extraction from VHR Remote Sensing Images" Sensors 23, no. 7: 3643. https://doi.org/10.3390/s23073643

APA Style

Wei, R., Fan, B., Wang, Y., & Yang, R. (2023). A Query-Based Network for Rural Homestead Extraction from VHR Remote Sensing Images. Sensors, 23(7), 3643. https://doi.org/10.3390/s23073643

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop