2.2.1. Field environment perception
The autonomous navigation of UAT requires environmental perception to perform tasks. The main objective of machine perception is to ensure that the UAT operates as expected and safely. Typical applications include obstacle detection, recognition of work area boundaries, and crop monitoring. Due to the challenging conditions faced by the sensors during agricultural operations, such as dust, rain, and extreme exposure to sunlight, sensor reliability has been a key challenge in environmental perception [
26]. Currently, various sensors are used in UATs for field environment perception, including monocular vision, stereoscopic vision, lasers, radar, and ultrasonic sensors.
(1) Visual perception
Machine vision localization, perception, and measurement are typically used for the UAT navigation at low speeds. Bakker et al. [
27] proposed a row recognition method based on the Hough transform. A color camera was used to capture images of sugar beets in a greenhouse. The color images were converted to grayscale images for creating sufficient contrast between the plant material and the soil background, substantially improving the image processing speed. To obtain real-time autonomous navigation information, Radcliffe et al. [
28] integrated a multispectral camera-based machine vision device on a small agricultural vehicle (
Figure 5a). The root mean square errors (RMSEs) for automatic navigation were 2.35 cm and 2.13 cm in laboratory and field environments, respectively.
A navigation algorithm of machine vision was developed for a rice field weeding robot [
29] (
Figure 5b). The results showed that the robot performed well at low weed density, with compensation accuracy of less than 2.5° and an average error from the target path of 4.59 cm. Mahboub and Mohammadi [
30] proposed a combined positioning method that integrated BDS and visual navigation, providing accurate and real-time obstacle information in agricultural fields. The position deviation of the tractor was within ±0.1 m, resulting in high accuracy of autonomous navigation. As shown in
Figure 5c, Ma et al. [
31] developed a visual module for a unmanned crawler tractor to obtain rice crop images in real time. The ExG(2G-R-B) algorithm, and the Otsu and mask method were used for segmenting the binary images.
Machine vision technology utilizes cameras as position measurement sensors. Image processing techniques are utilized to identify crop rows, determine a navigation reference line, and measure the relative position and heading information. The key advantages of machine vision technology are high speed, a large amount of information, and versatile functionality. However, the commonly used Hough transform algorithm has disadvantages, such as the difficulty in determining peak values, multiple repetitions of line segments, and high time and space complexity.
(2) Laser-based navigation
Thanpattranon et al. [
32] designed a control method for a tractor-trailer with single-sensor navigation system (
Figure 6a) used in orchards. A control scheme (
Figure 6a) for stopping the tractor-trailer using a laser range finder was designed for various tasks, as shown in
Figure 6b. The results demonstrated that the navigation of tractor in orchards had high accuracy, and the trailer position was adjusted by a sliding hitch bar, enabling wide turns in the paths between the trees. Laser navigation method has many strengths, such as high frequency, high accuracy, and large range. However, it has high costs. This technology is particularly suitable for agricultural robots.
(3) Inertial measurement unit
An inertial measurement unit (IMU) is a measurement instrument based on the principles of inertial navigation, usually consisting of three accelerometers and three gyroscopes. The integration of the angular velocity and acceleration data enables the estimation of the object's velocity, displacement, and attitude information, achieving navigation and positioning [
33]. Gyroscopes and accelerometers are the most common components of IMUs. In UATs, gyroscopes are used for autonomous navigation, operation control, and attitude measurement to improve operational efficiency and precision [
34]. An accelerometer consists of one or more acceleration sensors, and is widely used in inertial measurement systems of UATs. It has good bias stability and is resistant to vibrations, shocks, and temperature changes.
(4) Multi-sensor data fusion and perception
An INS is a closed-loop navigation system that does not have real-time external information to correct errors during motion. Thus, a single inertial navigation system can only be used for short-term navigation. Long-term navigation systems of UATs need satellite navigation to correct errors periodically.
Currently, multi-sensor data fusion is the most widely used approach for UAT navigation.
Figure 7 illustrates the combination of inertial and satellite navigation systems. Wang [
35] proposed a navigation method consisting of satellite/inertial navigation systems. Experiments were conducted on denoising the data from the satellite/inertial navigation systems, resulting in a navigation accuracy improvement of 2 m. Xia et al. [
36] combined information from an IMU and a GNSS. They used a robust regression approach to align the GNSS heading with the vehicle’s longitudinal motion. They also proposed a slip angle estimation method based on the dynamic model. The results showed improved estimation accuracy of the slip angle.
The Kalman filtering algorithm is the main method for sensor data fusion. It can reduce cumulative errors in inertial navigation. As shown in
Figure 8, Tian et al. [
37] developed a field robot integrated IMU and GNSS navigation system. Kalman filtering was used to correct the errors in the inclination data. Liu et al. [
38] proposed an integrated algorithm based on fuzzy reasoning and adaptive Kalman filtering for vehicle navigation and positioning using GPS and inertial navigation. The experimental results showed that the integrated algorithm had better positioning accuracy, precision, and stability than an RTK GPS positioning system.
Favorable results have been obtained from research and applications of laser radar and machine vision in the navigation and obstacle avoidance of UATs. Field environment perception enables the efficient implementation of tasks [
39]. However, there are still many challenges, such as data susceptibility to environmental interference, large data volumes, and insufficient real-time performance. Further efforts are needed to enhance the robustness of detection algorithms.
2.2.2. Operation state perception
The perception of operational conditions of UATs primarily focuses on the engine, steering system, transmission system, vehicle body, and wheels. Key parameters include torque, rotational speed, emissions, attitude, vehicle speed, wheel speed, vibration, and strain [
40]. The operational condition information includes the traction force of the suspension system, suspension lifting position, power take-off (PTO) torque, PTO speed, hydraulic flow rate, and pressure.
Table 1 lists the perception methods for the operational states of UATs.
A subsoiler equipped with flexible tines allows for obstacle avoidance while minimizing draft force. However, due to the substantial variation in soil resistance, tilling often results in depths that are considerably lower than the desired target value. To address this issue, researchers developed an electric-hydraulic system for a subsoiler as depicted in
Figure 9a [
46]. Additionally, they introduced a novel method for detecting the tillage depth to overcome this challenge. The results showed that the control system improved the tillage quality of the subsoiler with flexible tines. Wang et al. [
47] devised a precise perception system (
Figure 9b) for corn fertilization planters. A capacitance sensor was designed to detect the amount of fertilizer online based on the different dielectric properties of fertilizer and air. The electrically driven seed metering system exhibited an impressive control accuracy of 98% for controlling the grain spacing.