Height-Variable Monocular Vision Ranging Technology For Smart Agriculture
Height-Variable Monocular Vision Ranging Technology For Smart Agriculture
Height-Variable Monocular Vision Ranging Technology For Smart Agriculture
ABSTRACT Smart agriculture utilizes a variety of advanced technologies to promote sustainable agriculture
and provide solutions for intelligent, automated and unmanned agriculture. Agricultural robots and related
technologies are an important part of smart agriculture, while autonomous navigation is a core function
of autonomous agricultural robots, which rely on information about the distance of obstacles in a scene
to support decision making. In this paper, we propose a ground point geometric ranging model, which
can be used in camera height dynamic change scenarios, and the method is validated by model derivation
and hypothesis testing. The model combines ranging and camera calibration, choosing to compensate for
distortion and defocus phenomena caused by nonlinear imaging of the camera to the focal length, and
completes the parameter calibration using a small amount of ground point real distance data. In this paper,
the YOLOv8 model is used to identify and range outdoor cattle, and the experimental results show that the
lowest range accuracy of this method reaches 95%, this method eliminates the dependence on camera height
for focal calibration in ranging models, and in practice requires only once focal calibration for permanent
use, achieving a significant reduction in the complexity of focal calibration, and the migrability of the model
in scenarios where the camera height changes.
INDEX TERMS Monocular vision, distance measurement, focal length calibration, smart agriculture,
YOLOv8.
In smart agriculture, distance measurement also has impor- during inference may significantly increase the efficiency
tant research value in areas such as automatic driving of and the performance, the network requirement for labeled
agricultural vehicles and terrain mapping of farmland. ground truth depth data decreased. Masoumian et al. [11]
The purpose of this paper is to investigate ranging tech- proposed a Unsupervised Learning Approach, they devel-
niques that can be applied in low-cost agriculture, achieving oped a multi-scale Monocular Depth Estimation based on a
the ranging of objects with contact points with the ground, graph convolutional network. Their network consists of two
such as livestock and crops. Based on the team’s previous parallel autoencoder networks: DepthNet and PoseNet. The
research results, we propose a ground point geometric rang- DepthNet is an autoencoder composed of two parts: encoder
ing model, which can be used in camera height dynamic and decoder; the CNN encoder extracts the feature from the
change scenarios, and the method is validated by model input image, and a multi-scale GCN decoder estimates the
derivation and hypothesis testing. The model combines rang- depth map. PoseNet is used to estimate the ego-motion vector
ing and camera calibration, choosing to compensate for dis- between two consecutive frames. The estimated 3D pose and
tortion and defocus phenomena caused by nonlinear imaging depth map are used to construct a target image. This approach
of the camera to the focal length, and completes the parameter has a high prediction accuracy of 89% on the publicly KITTI
calibration using a small amount of ground point real distance and Make3D datasets, with a 40% reduction in the number
data. In this paper, the YOLOv8 model is used to identify and of trainable parameters. Compared with deep learning-based
range outdoor cattle, and the experimental results show that ranging methods, geometric model-based ranging methods
the lowest range accuracy of this method reaches 95%. require fewer training parameters and less ground truth depth
data, and the scenes are more transferable.
II. LITERATURE SURVEY Camera calibration is the basis of visual ranging, and
Commonly used distance measurement methods include the geometric model-based ranging methods need to obtain
radar distance measurement [5], laser distance measure- the calibration parameters of the camera first, so the main
ment [6], and visual distance measurement. The radar ranging calibration methods of the camera need to be introduced
system has high accuracy, but the equipment is costly and here before introducing the geometric-based ranging model.
vulnerable to weather; the laser ranging system is low cost, The main purpose of camera calibration is to determine
but requires a clean environment free of foreign objects such the relationship between the 3D information of the surface
as dust; the visual ranging system is low cost, has no special points of the object and the corresponding image points by
requirements for the use environment, and is more suitable for constructing a linear imaging model of the camera. There are
use in the complex outdoor environment of farms than other three methods of camera calibration, namely the traditional
methods. camera calibration method, the active vision camera calibra-
Vision ranging systems can be classified into monocular tion method, and the camera self-calibration method [12],
systems and stereo systems [7]. In a stereo system it is [13]. The traditional camera calibration method is applica-
necessary to match the pictures captured by multiple cam- ble to any camera and has high accuracy, but requires a
eras. When the distance from the camera to the scene is calibration object and a complex algorithm. The traditional
greater than the baseline of the stereo camera (the distance camera calibration method proposed by Zhang [14] uses a
between two cameras), the stereo vision degrades to monoc- calibration plate composed of two-dimensional squares for
ular vision. The monocular vision system does not need to calibration, acquires pictures of different poses of the cali-
match the image in the data preprocessing stage and the bration plate, extracts the pixel coordinates of corner points
hardware cost is low [8], so it is a good choice to obtain the in the pictures, calculates the initial values of the internal
depth information of the image target. At present, the com- and external parameters of the camera through homography
monly used monocular vision ranging methods include deep matrix, estimates the distortion coefficients using nonlinear
learning-based ranging methods and geometric model-based least squares, and finally optimizes the parameters using
ranging methods. In the deep learning-based ranging meth- the maximum likelihood estimation method. Li et al. [15]
ods, Qi et al. [9] proposed a Supervised Learning Approach, presented a calibration method that uses feature extraction
this method utilized two networks to estimate the depth map techniques to encode feature points that contain features at
and surface normal from single images. These two networks many different scales. It can be used for internal and external
enable the conversion of depth-to-normal and normal-to- calibration of multi-camera systems, as well as for internal
depth and collaboratively increase the accuracy of the depth calibration of individual cameras. The active vision-based
map and surface normal. Although their neural network can camera calibration method [16] is based on an active system
increase the accuracy of depth maps, for training, they require that controls the camera to make a specific motion and take
ground truth, including surface normal, which is hard to multiple sets of images, and solves the internal and external
obtain. Luo et al. [10] proposed a Semi-Supervised Learning parameters of the camera based on the image information and
Approach, this method showed that the monocular depth known displacement changes. This calibration method does
estimation problem can be reformulated as two sub-problems, not require calibration object, and the algorithm is simple and
a view synthesis procedure followed by stereo matching. they robust, but it needs to be equipped with an accurate control
corroborated that the application of geometric limitations platform, so the cost is high. The self-calibration method
based on Kruppa [17] establishes the constraint equation point on the same ray passing through the camera optical
about the camera’s internal parameter matrix through the center has the same pixel coordinates, it is guessed that the
quadratic curve, and uses at least 3 pairs of images to calibrate focal length value of the pixel point is independent of the
the camera. The length of the image sequence will affect the camera height. In this paper, we will demonstrate through
stability of the calibration algorithm, and the infinite plane theoretical deduction and hypothesis testing that under the
in the projective space cannot be guaranteed. The camera calibration method proposed by Xue, the change of camera
self-calibration method does not need the calibration object, height will not affect the focal length of the pixel point,
and it has strong flexibility and can be calibrated online, but that is, the calibration result of focal length can be applied
the accuracy is low and the robustness is poor. at full camera height with a fixed camera tilt angle, thus
In the geometry-based ranging model, the method pro- achieving the purpose of reducing the number of focal length
posed by Mao et al. [18] measures the distance of an object calibrations and expanding the use scenario of the ranging
that has a contact point with the ground and the height of method. The main innovations of this study are as follows:
the object to be measured is known. The exact distance 1) A geometric ranging model combining camera calibration
can be obtained based on similar triangles when the vertex is presented, and it is demonstrated through model derivation
imaging is located in the center of the image plane, and when and hypothesis testing that the method can be permanently
the vertex imaging is not in the center of the image plane, available with a single calibration. This allows the model to
the approximate solution of the distance is obtained. The be used in ranging scenarios where the camera height varies
disadvantage of this method is that the object needs to be arbitrarily. 2) Combining the latest YOLOv8 target detection
perpendicular to the ground, and the approximate solution model with the ranging model proposed in this paper, a lowest
can show higher accuracy when the object distance is much range accuracy of more than 95% is achieved in cattle ranging
larger than the camera height, or the camera height is much on outdoor farms.
larger than the object height. The ranging model proposed
by Martínez-Díaz [19] does not need to consider the camera III. MATERIALS AND METHODS
pose, but needs to know the relative distance between the A. ESTABLISH GEOMETRIC MODEL AND INFERENCE
three points in the image including the measured point, so the PROOF
application scenario is limited. Most of the current ranging In the geometric model proposed by Xue, the distance from
methods based on geometric ranging models do not take the point on the plane to the camera is:
into account the impact of camera nonlinear imaging on the
H · x 2 + f 2 − y · f · tan α
model, and require known spatial information such as the
d= p (1)
height of the object being measured or the distance between x 2 + f 2 · (f · tan α + y)
feature points during the ranging process.
The calibration method proposed by Xue et al. [20] links As shown in (1), H is the height of the camera, x and y
the geometric ranging model and the focal length calibration are the physical coordinates of the measured point, and α is
together, differs from the previous calibration idea of correct- the downward tilt of the camera. The focal length f of the
ing the pixel coordinates with distortion. In this method, the pixel point which corresponding the measured point in the
distortion and defocusing phenomena caused by the nonlinear image can be found when d is known. This calibration method
imaging of the camera are innovatively compensated to the compensates the distortion and defocusing caused by the
focal length of each pixel point, so the focal length of the non-linear imaging of the camera to the focal length, so the
camera is not considered as a constant value in this calibration focal length of each pixel point is not the same. This can be
method, and the expression of the focal length of a pixel explained that this aberration reduction method considers that
point with its physical coordinates needs to be obtained using the light reflected from the measured point remains linearly
polynomial regression. The paper proposes a ranging model imaged after passing through the optical centre, but that the
when the camera has tilt in three dimensions, and experiments different imaging points correspond to different positions
prove that the ranging accuracy under this calibration method of the imaging plane, and therefore different focal length
reaches more than 97%. In the discussion part of the article, values f for different imaging points. The focal length at
Xue uses experimental data to point out that the focal length point P1 is noted as f1 and the focal length at point P2 is
regression vector v calculated by the camera at different noted as f2.
heights is not the same, but does not further investigate the In Fig.1, point O is the location of the camera, and the
effect of camera height change on the focal length of pixel distances of point O from plane 1 and plane 2 are H1
points. Therefore, it needs to be calibrated separately for and H2. P1 and P2 are the intersections of the rays pass-
each installation height, which cannot meet the use in height ing through point O with plane 1 and plane 2. Since P1
change scenarios. and P2 are in the same ray passing through the optical
In a word, this paper is a follow-up study to the cali- center, the projection points of P1 and P2 on the image
bration method of Xue, and aims to illustrate the effect of plane are the same, both being P′ (P′ X, P′ Y). d1 and d2
camera installation height on the focal length calibration are the distances of points P1 and P2 from the camera,
results of pixel points in this method. Since any measured respectively.
B. HYPOTHESIS TESTING
Hypothesis testing, also known as statistical hypothesis test-
ing, is a statistical inference method used to judge whether
differences between samples and samples, or between sam-
ples and the population, are caused by sampling error or
by essential differences. Significance testing is one of the
most commonly used methods in hypothesis testing, and it
is also the most basic form of statistical inference. The basic
principle is to first make a certain assumption about the
characteristics of the population, and then make an inference
about whether this assumption should be rejected or accepted
through the statistical inference of sampling research. Com-
monly used hypothesis testing methods include Z-Test, T-
Test, Chi-square Test, F-Test, etc.
In this paper, the focal length values of multiple pixel
points at different heights are obtained by shooting infor-
mation, and then the relationship between the focal length
values of the same pixel points at different camera heights
is investigated by Paired-Samples T Test. The basic steps
are as follows: (a) The original hypothesis is proposed as:
FIGURE 1. Ranging models at different camera heights. the population mean of the focal length of the pixel points
at different heights has no significant difference, expressed
Since d1 // d2, OO1P1∽OO2P2, according to the proper- as H0 = µ1 − µ2. µ1 and µ2 are the population mean
ties of similar triangles can be obtained: values of the two paired samples, respectively. (b) Construct
H 1· P′2 +f 12 −P′Y ·f 1·tan α
the statistics:
q X
X +f 1 ·(f 1·tan α+PY )
P′2 2 ′
d1
=
OO1
=
H1
= (2) d̄ − (µ1 − µ2 )
d2 OO2 H2 H 2· P′2 +f 22 −P′Y ·f 2·tan α
t= s√ (4)
q X n
X +f 2 ·(f 2·tan α+PY )
P′2 2 ′
After the (2) is reduced, can be obtained, so the focal d̄ is the mean of the difference between the two paired sam-
lengths of P1 and P2 are equal. ples and µ1 − µ2 is the difference between the mean values
According to the above proof, the focal length values and of the two populations. (c) Calculation of the observed values
physical coordinates of P1 and P2 are equal. Since light of the test statistic and the corresponding probability P. (d)
travels in a straight line, when light intersects a plane in Given the significance level α, compare it with the probability
space, it must also have an intersection point in another plane P of the test statistic. If the probability P is less than the
parallel to that plane. So for any point P1 on plane 1 you can significant level α, the original hypothesis should be rejected
find a point P2 on plane 2 with the same focal length value and the mean values of the two populations are significantly
and physical coordinates as P1. different, that is, the transformation of camera height has a
In order to obtain the focal length values of different significant impact on the focal length of pixel point. On the
pixel points, Xue uses polynomial regression to learn the contrary, if the probability P is greater than the significant
relationship between the focal length f of a pixel point and level α, the original hypothesis should not be rejected and
the physical coordinates (x,y). The physical coordinate is the there is no significant difference between the two population
independent variable and the focal length is the dependent means, that is, the transformation of the camera height has no
variable, (3) and regression vector V are obtained through significant effect on the focal length of pixel point. In order to
polynomial regression. reflect the reliability of the conclusion from a statistical point
of view, the experiment was conducted using SPSS software
f (x, y) = v00 + v10 · x + v01 · y + v11 · x · y + v00 · y2 · · · to perform Paired-Samples T Test on multiple sets of focal
(3) length data. The original hypothesis was that there was no
And because the regression vector V in this calibration significant difference between the two sets of focal length
method is only related to the physical coordinates and focal data at different heights, the significance level is set at 95%.
length value of the pixel point, and any parallel plane can This experiment uses a 1920∗ 1080 pixel zoom camera, the
find the pixel point with the same physical coordinate and camera attitude is set as downward inclination α = 35◦ and
focal length value, so the same set of regression vectors can left inclination β = 0◦ , five camera heights are randomly
be applied to parallel planes. It can be proved that in Xue’s selected as H1 = 0.515m, H2 = 0.568m, H3 = 0.646m,
calibration method, the calibration result of the focal length H4 = 0.756m, and H5 = 0.809m. The experiment obtains
is not affected by the height transformation of the camera. the shooting information of 30 obstacles at different cam-
C. RANGING VERIFICATION
In order to reflect the reliability of the conclusion from els to improve the model performance. 3. Head: The Head
the distance measurement accuracy, the regression vectors part is replaced by the current mainstream Decoupled-Head
V under different heights of H1- H5 are first obtained by structure, which separates the classification and detection
polynomial regression, which are noted as V1- V5, and head, and also changed from Anchor-Based to Anchor-Free.
then calculate the obstacle distance obtained by using the 4. Loss: YOLOv8 abandoned the previous IOU matching
regression vector corresponding to the height and the obstacle or unilateral proportional distribution method, using the
distance obtained by using the regression vector V3, and Task-Aligned Assigner positive and negative sample match-
finally evaluate the distance error obtained by the two meth- ing method, and the introduction of Distribution Focal Loss
ods. Experiment uses MATLAB2022 to calculate the focal (DFL). 5. Train: The data enhancement part of the training
length value, focal length regression vector and distance to introduces the last 10 epoch off Mosiac enhancement opera-
the object. This experiment uses the error metric function tion in YOLOX, which effectively improves the accuracy.
Fig.3 shows a schematic of the network structure of
1 Xn Real (t) − Pred (t) YOLOv8. Fig.4 shows the complete ranging process and
MAPE = (5)
n t Real (t) calibration method, and includes the comparison before and
as the error evaluation index, because MAPE considers not after the improvement.
only the error between the predicted value and the true value, A zoom camera with 1920∗ 1080 pixels was used for the
but also the ratio between the error and the true value, so it is experiment, and the camera pose was set to downward tilt.
suitable for the evaluation of ranging accuracy. The experiments used a laser rangefinder to obtain the actual
Our method is planned for use in a low-cost agricul- distance of the measured point, which was used to verify the
tural application scenario, aiming to measure the distance of accuracy of the range finding results. The experimental steps
objects such as livestock and crops that have a point of contact are as follows: 1) calibrate the focal length of the image plane
with the ground, enabling intelligent grazing by unmanned at the camera height of 1.40m, and use the focal length regres-
farming machines. In order to verify the transferability of sion vector for the physical coordinates of all imaging points.
the ranging model in a real-world scenario, we conducted 2) adjust the camera height to 1.50m and acquire the images
experiments on an outdoor farm. of the measured points. 3) output the distance measurement
In practice, we need to identify the object to be results, compare them with the distance acquired by the laser
measured, and then complete the distance measure- rangefinder and output the distance measurement accuracy.
ment based on its output position information. This The experiment uses MATLAB2022 to calculate the focal
paper uses the official YOLOv8 open source code length value of the measured point, and uses python 3.9 for
(https://github.com/ultralytics/ultralytics) for target detec- cattle identification and ranging.
tion. YOLOv8, a major update published by ultralytics on The pre-trained YOLOv8 model downloaded from the offi-
January 10, 2023, now supports image classification, object cial website, as well as includes the recognition of cattle,
detection and real-time segmentation tasks. so we directly input the subject image into the pre-trained
The core features and changes of the YOLOv8 algorithm YOLOv8 model for target detection and position extraction.
are summarized as follows: 1. A new SOTA model is pro-
vided, including P5 640 and P6 1280 resolution target detec- IV. RESULTS
tion networks and YOLACT-based instance segmentation A. FOCAL LENGTH VALUE HYPOTHESIS TESTING RESULTS
models. Different size models based on scaling factors are F(H1) - F(H5) in Fig.5 is the focal length of 30 pixels points
also provided for different scenarios at N/S/M/L/X scales. obtained by using (1) when the camera height is H1∼H5,
2. Backbone: Backbone and Neck sections replace the C3 Fig.6 shows the regression models at five different heights.
structure of YOLOv5 with a more gradient-rich C2f structure, The Paired-Samples T Test results are shown in Table 1. Since
and adjust the number of channels for different scale mod- the probability P for all paired tests are greater than 0.05,
V. DISCUSSION
A. PIXEL POINTS OF MULTIPLE CAMERA HEIGHTS
JOINTLY CALIBRATE THE FOCAL LENGTH
In Sec. III-A, it is demonstrated that the calibration results of
the pixel point focal length are not affected by the camera
height transformation, which not only means that the cali-
bration of the pixel point focal length can be done at only
one camera height when the camera tilt angle is constant, but
also shows that the calibration results are not affected when
FIGURE 4. Ranging process and calibration method. using pixel points from multiple camera heights together to
calibrate the focal length. In order to verify the correctness of
it indicates that the mean difference between the pixel point this conclusion, this paper first calibrates the focal length of
focal lengths obtained at height H3 and those obtained at the camera using the shooting information of 30 pixel points
other heights is not statistically significant compared to 0. at the camera height of 0.646m, and the calibration result
is recorded as F(fixH); then calibrates the focal length of
B. RANGING VERIFICATION RESULTS the camera using the shooting information of these 30 pixel
Table 2 shows the shooting information of two randomly points at five different camera heights, and the calibration
selected pixel points, where (x,y) is the pixel coordinates of result is recorded as F(mixH). Paired-Samples T Test was
TABLE 3. Range results for the four cows (labeled from left to right). TABLE 5. Focal length calibration and ranging at single camera height
and multiple camera heights.
[3] A. Abdalla, H. Cen, L. Wan, K. Mehmood, and Y. He, ‘‘Nutrient sta- MEIAN LI was born in Dazhu, Sichuan, China,
tus diagnosis of infield oilseed rape via deep learning-enabled dynamic in 1973. He received the Ph.D. degree in computer
model,’’ IEEE Trans. Ind. Informat., vol. 17, no. 6, pp. 4379–4389, systems architecture from the University of Elec-
Jun. 2021. tronic Science and Technology, in 2007.
[4] O. E. Apolo-Apolo, J. Martínez-Guanter, G. Egea, P. Raja, and He has presided over one emergency project
M. Pérez-Ruiz, ‘‘Deep learning techniques for estimation of the yield and of the National Natural Foundation of China, one
size of citrus fruits using a UAV,’’ Eur. J. Agronomy, vol. 115, Apr. 2020, project of the Inner Mongolia Natural Foundation
Art. no. 126030.
of China, and participated in two projects of the
[5] L. Piotrowsky, S. Kueppers, T. Jaeschke, and N. Pohl, ‘‘Distance mea-
National Natural Foundation of China and two
surement using mmWave radar: Micron accuracy at medium range,’’ IEEE
Trans. Microw. Theory Techn., vol. 70, no. 11, pp. 5259–5270, Nov. 2022. projects of the Inner Mongolia Natural Foundation
[6] L. Lombardi, V. Annovazzi-Lodi, G. Aromataris, and A. Scirè, ‘‘Distance of China. He has published more than 30 articles, including nearly 20 SCI and
measurement by delayed optical feedback in a ring laser,’’ Opt. Quantum EI retrieved articles. His current research interests include data intelligence
Electron., vol. 54, no. 5, p. 270, May 2022. and software engineering, particularly theoretical methods and applications
[7] A. P. Nugroho, M. A. N. Fadilah, A. Wiratmoko, Y. A. Azis, A. W. Efendi, of vehicle monocular vision.
L. Sutiarso, and T. Okayasu, ‘‘Implementation of crop growth moni-
toring system based on depth perception using stereo camera in plant
factory,’’ IOP Conf. Ser., Earth Environ. Sci., vol. 542, no. 1, Jul. 2020, LIXIA XUE was born in Ulanqab, China, in 1997.
Art. no. 012068. She received the bachelor’s degree in informa-
[8] P. Ferrara, A. Piva, F. Argenti, J. Kusuno, M. Niccolini, M. Ragaglia, tion management and information system and the
and F. Uccheddu, ‘‘Wide-angle and long-range real time pose estimation: master’s degree in computer application technol-
A comparison between monocular and stereo vision systems,’’ J. Vis. ogy from Inner Mongolia Agricultural University
Commun. Image Represent., vol. 48, pp. 159–168, Oct. 2017. in 2019 and 2022, respectively. Currently, she is
[9] X. Qi, R. Liao, Z. Liu, R. Urtasun, and J. Jia, ‘‘GeoNet: Geometric working as a Project Manager in Inner Mongolia
neural network for joint depth and surface normal estimation,’’ in Proc. Power Group Mengdian Information and Commu-
IEEE/CVF Conf. Comput. Vis. Pattern Recognit., Jun. 2018, pp. 283–291. nication Industry Company. Her research interests
[10] Y. Luo, J. Ren, M. Lin, J. Pang, W. Sun, H. Li, and L. Lin, ‘‘Single include computer vision, unmanned driving, etc.
view stereo matching,’’ in Proc. IEEE/CVF Conf. Comput. Vis. Pattern
Recognit., Jun. 2018, pp. 155–163.
[11] A. Masoumian, H. A. Rashwan, S. Abdulwahab, J. Cristiano, M. S. Asif, JINGWEN BAO was born in China, in 1998. She
and D. Puig, ‘‘GCNDepth: Self-supervised monocular depth estima- majored in information management and infor-
tion based on graph convolutional network,’’ Neurocomputing, vol. 517, mation systems at Inner Mongolia Agricultural
pp. 81–92, Jan. 2023, doi: 10.1016/j.neucom.2022.10.073. University. She received management degree in
[12] Q.-T. Luong and O. D. Faugeras, ‘‘Self-calibration of a moving camera 2021. During her graduate studies, she received the
from point correspondences and fundamental matrices,’’ Int. J. Comput. master’s degree in electronic information in 2023.
Vis., vol. 22, no. 3, pp. 261–289, Mar. 1997. Her research focuses on computer vision.
[13] W. Dong and V. Isler, ‘‘A novel method for the extrinsic calibration of
a 2D laser rangefinder and a camera,’’ IEEE Sensors J., vol. 18, no. 10,
pp. 4200–4211, May 2018.
[14] Z. Zhang, ‘‘A flexible new technique for camera calibration,’’ IEEE Trans.
Pattern Anal. Mach. Intell., vol. 22, no. 11, pp. 1330–1334, Nov. 2000.
HAO LIAN was born in Ulanqab, China, in 2000.
[15] B. Li, L. Heng, K. Koser, and M. Pollefeys, ‘‘A multiple-camera system
He received the Bachelor of Engineering degree
calibration toolbox using a feature descriptor-based calibration pattern,’’ in
in 2022. He currently holds the master’s degree
Proc. IEEE/RSJ Int. Conf. Intell. Robots Syst., Nov. 2013, pp. 1301–1307.
[16] Y. Xu, F. Gao, Z. Zhang, and X. Jiang, ‘‘A calibration method for non- from Inner Mongolia Agricultural University. His
overlapping cameras based on mirrored absolute phase target,’’ Int. J. Adv. research interests include computer vision and
Manuf. Technol., vol. 104, nos. 1–4, pp. 9–15, Sep. 2019. deep learning.
[17] J. Li, Y. Yang, and G. Fu, ‘‘Camera self-calibration method based on GA-
PSO algorithm,’’ in Proc. IEEE Int. Conf. Cloud Comput. Intell. Syst.,
Sep. 2011, pp. 149–152.
[18] M. Jiafa, H. Wei, and S. Weiguo, ‘‘Target distance measurement
method using monocular vision,’’ IET Image Process., vol. 14, no. 13,
pp. 3181–3187, Nov. 2020. TIN LI was born in China, in 1999. She holds the
[19] S. Martínez-Díaz, ‘‘3D distance measurement from a camera to a mobile bachelor’s degree in information management and
vehicle, using monocular vision,’’ J. Sensors, vol. 2021, pp. 1–8, Apr. 2021. information systems big data from Inner Mongolia
[20] L. Xue, M. Li, L. Fan, A. Sun, and T. Gao, ‘‘Monocular vision ranging Agricultural University, where is currently pursu-
and camera focal length calibration,’’ Sci. Program., vol. 2021, pp. 1–15, ing the master’s degree in computer technology.
Jul. 2021. Her research direction focuses on the field of com-
puter vision.
TIAN GAO was born in Ulanqab, China, in 1997. YANYU SHI was born in Ulanqab, China, in 1999.
She received the B.S. degree in information and He received the bachelor’s degree in computer
computing science from Inner Mongolia Nor- science and technology from Inner Mongolia Agri-
mal University in 2019, where she pursued her cultural University in 2022, where he is currently
M.S. degree in computer application technology in pursuing the master’s degree in computer tech-
2020 to 2023, respectively. Her research interests nology. His research interests including intelligent
include computer vision ranging, smart agricul- animal husbandry and computer vision.
ture, etc.