Skip to main content
Sensors (Basel, Switzerland) logoLink to Sensors (Basel, Switzerland)
. 2021 Jul 30;21(15):5168. doi: 10.3390/s21155168

Towards an Optimal Footprint Based Area Coverage Strategy for a False-Ceiling Inspection Robot

Thejus Pathmakumar 1, Vinu Sivanantham 1, Saurav Ghante Anantha Padmanabha 1, Mohan Rajesh Elara 1,*, Thein Than Tun 2
Editor: Alberto Borboni
PMCID: PMC8347183  PMID: 34372408

Abstract

False-ceiling inspection is a critical factor in pest-control management within a built infrastructure. Conventionally, the false-ceiling inspection is done manually, which is time-consuming and unsafe. A lightweight robot is considered a good solution for automated false-ceiling inspection. However, due to the constraints imposed by less load carrying capacity and brittleness of false ceilings, the inspection robots cannot rely upon heavy batteries, sensors, and computation payloads for enhancing task performance. Hence, the strategy for inspection has to ensure efficiency and best performance. This work presents an optimal functional footprint approach for the robot to maximize the efficiency of an inspection task. With a conventional footprint approach in path planning, complete coverage inspection may become inefficient. In this work, the camera installation parameters are considered as the footprint defining parameters for the false ceiling inspection. An evolutionary algorithm-based multi-objective optimization framework is utilized to derive the optimal robot footprint by minimizing the area missed and path-length taken for the inspection task. The effectiveness of the proposed approach is analyzed using numerical simulations. The results are validated on an in-house developed false-ceiling inspection robot—Raptor—by experiment trials on a false-ceiling test-bed.

Keywords: functional footprint, false-ceiling inspection, multi-objective optimization, pest-control robot, path-planning

1. Introduction

Pest management is one of the critical elements in facility management that has shown growth to USD 22.7 billion in 2021 and is estimated to reach USD 29.1 billion by 2026 globally [1,2]. The Rodent Monitoring System (RMS) is one of the necessary tasks in a facility management system since the rodents are a critical transmission vector for various diseases [3,4,5,6]. In addition, the infestation of rodents brings loss of revenue from damaged appliances, structural damages, loss of commodities, and expenses for involving pest-control services and professionals [7,8,9].

There have been a variety of rodent inspection techniques developed over time to check for any rodent infestation. An Internet of Things (IoT)-based rodent infestation monitoring system is reported in [10], where the authors utilized sensor-based and vision-based algorithms for rodent monitoring. Ross et al. reported the development of a smart rodent monitoring system called RatSpy that uses computer vision and IoT for inspection within a rodent bait station [11,12]. Sowmika et al. discuss a method of rodent detection in farmlands, where a PIR sensor is used to detect the motion of rodents within a range of ten meters, and sensor data collected are pushed to the cloud for enabling remote monitoring [13]. Similarly, the work mentioned in [14] uses wireless sensor networking for rodent detection. The research outcome disclosed in [15] describes the development of rodent bait equipped with electronic sensors in its bed to detect the presence of rodents within the bait. The prior-art study shows that sensing techniques, IoT, and computer vision are catalysts for the proliferation of rodent inspection techniques. However, the advancements in IoT-based techniques for rodent management and infestation control cannot be considered a perfect solution. The omnipresent rodent species, such as Norway rats (Rattusnorvegicus) and black rats (Rattusrattus), and their strong and consistent behavioral characteristics, such as neophobia [16,17], increase their awareness of trap locations, and so they keep changing their harborage. Furthermore, it is extremely challenging to locate a dead and decaying rodent inside a false ceiling after rodenticide consumption. The scenario mentioned above gives rise to the need to formulate effective methods for rodent activity monitoring inside a false ceiling. In addition to the rodent inspection, another important task to be carried out is the inspection of building structures like optic-fiber cables, sprinkler and gas pipes, and air-condition vents prone to rodent gnaw bites as part of the periodic maintenance. The conventional method for the inspection task is a manual inspection done by humans. This method is ineffective since the person who inspects the false ceiling gets an eclipsed vision by obstacles most of the time. Hence, the manual inspection methods are inefficient and ineffective in terms of time consumption and inspection quality. Furthermore, the risk factor involved in climbing up to high-rise ceilings multiple times escalates the drawbacks of manual false ceiling inspection.

A vast volume of research works is reported on the development of strategies for automated inspection tasks for complex environments [18,19,20,21]. Robot-aided inspection is one of the key factors among automated inspection, which has been widely researched recently. For instance, work mentioned in [22] reports the development effort of flexible pipe inspection robots, where authors propose a bio-inspired caterpillar-like robot for piping inspection. Similarly, Imran et al. developed an inspection robot capable of passing through complicated pipelines with full control [23]. Similar to pipeline inspection, autonomous robots are being used for handling high-risk inspection tasks inside nuclear reactors, where authors propose a lightweight robot with a holonomic base and a manipulator [24]. The development of a novel drainage inspection robot is mentioned in [25], where authors propose strategies for the collision avoidance and stability control of the drainage inspection robot. Additionally, an inspection of contacts on overhead cable systems for high-speed rails using inspection robots is discussed in [26]. Leveraging on the precedence on effective usage of robots for complex inspection strategies and the need to eliminate risk and inefficiency factors involved in false-ceiling inspection gives rise to a new modality of inspection, which is the usage of lightweight inspection robots. A robot that is capable of navigating around the false ceiling either in an autonomous or semi-autonomous fashion can perform rodent activity monitoring and periodic false ceiling inspection efficiently. However, the brittle nature of false-ceiling panels installed inside a building restricts heavy structural weight. To maximize accessibility, the dimensions of the robot have to be compact. The design considerations mentioned above on structural weight and dimensions do not allow the robot to carry a heavy battery on board. Hence, the inspection coverage planning has to be efficient in terms of energy consumption.

Autonomous navigation is an essential factor for autonomous robots. There is numerous research work reported on perception, navigation, and localization methods [27,28,29,30,31]. Among all, a significant research focus is given on the path planning strategy since it can have an impact on the task performed by the robot. For instance, the work mentioned in [32] describes the path-planning approach in a mobile robot for a gas mapping task. Miao et al. proposed a scalable coverage planning strategy for a cleaning robot [33]. From the author’s perspective, a path-planning approach for area coverage is a critical element in robotic cleaning. Sasaki et al. proposed an adaptive path planning strategy for improving the performance of a cleaning robot, where the coverage planning prioritizes and sorts regions based on dirt distribution [34]. The footprint of the robot is a crucial element of area coverage algorithms. A classic robot footprint is defined as the surface area occupied by the robot at any instant of time [35]. For complete coverage planning, the robot should be navigating so that the defined robot footprint will cover the maximum possible area. The same strategy is applied for complete visual coverage tasks as well [36]. Hence, for a false ceiling inspection, the effectiveness of the task depends on the robot’s ability to cover every location in a false ceiling. For a visual inspection, the robot footprint can be defined as the maximum area within the camera’s field of view that can be inspected. However, the efficiency of inspection also has an impact on the choice of an optimal footprint.

The robot footprint used to maximize the task efficiency is also defined as the functional footprint of the robot [37]. In this paper, we define an optimal footprint, also known as functional footprint, for an in-house developed false ceiling inspection robot called Raptor. In the scenario mentioned above, the robot’s functional footprint depends upon the position and orientation of the camera onboard. We have exploited genetic algorithms to determine the functional footprint of the robot that ensures an efficient false-ceiling inspection. The rest of the research article is organized as follows:

Section 2 details the general objective of this work, followed by the system architecture of the Raptor robot in Section 3. Section 4 explains the coverage planning strategies used by the Raptor robot, and Section 5 gives a detailed insight regarding the cost functions, constraints, and optimization strategies for finding the functional footprint. The results of the simulation are given in Section 6, and finally, the conclusion and future works are given in Section 7.

2. Objectives

The functional footprint of the robot is the footprint area of the robot that enhances the efficiency of task execution [37]. The general objective of this work is to define a functional footprint for the Raptor robot and improve the performance of the inspection task the robot carries out by considering the defined functional footprint. This general objective is subdivided into three components:

  • A zig-zag area coverage strategy is proposed for false-ceiling inspection, where the area coverage is defined in terms of the robot footprint, which is dependent on the position and orientation of the camera (camera installation parameters).

  • Formulate cost functions based on area missed and path-length covered during the inspection by the robot.

  • Estimate the functional footprint by finding the optimal values for camera installation parameters by minimizing the costs using multi-objective optimization.

  • Validate the results of the Raptor robot by conducting experiment trials on a false ceiling test-bed.

3. System Architecture

The outline of the in-house developed Raptor robot and its associated components are given in Figure 1. The robot was designed by considering the maximum supported weight on a false ceiling of 2.0 kg. The overall weight of the robot with a 2800 mAH LiPo battery is 1.2 kg, and the weight without a battery is approximately 1.03 kg. The developed robot system comprises four main segments: Operator’s Console Unit, ROS master controller, Payload platform, and Beacon platform. The ROS master control unit controls the movement of all deployed robots at any given time. There are two variants for the Raptor robot platform, namely: Payload platform and Beacon platform. The payload platform can perform semi-autonomous navigation using the RPlidar A1M8 LIDAR sensor as the primary sensor for localization and navigation. On the other hand, the beacon platform works with UWB-based mobile and stationary beacons for indoor localization. The Operator’s Console Unit is the Graphic User Interface (GUI) layer to interface with the ROS master. The ROS master synchronizes the platform data between the beacon platforms and the payload platforms. The platform has the dimensions: 390×350×200 mm. We will focus on the Beacon platform for the false ceiling inspection task since we require map-less and GPS-denied indoor navigation. The beacon platform has a skid-steering base powered by four DC Motors that operate on a closed-loop velocity control. A pair of Roboclaw motor controllers drive the motor according to the high-level command control received from a single board computer. Raspberry Pi-4 Model B is used as the slave single-board computer for processing of high-level computation, communication, and control with the remote ROS master controller. The motor driver and Raspberry Pi controller are placed inside a unique chassis to avoid interference induced by the payload and DC motors. One Roboclaw controls the front two motored wheels while another drives the back two motored wheels. Each Roboclaw is connected to a pair of motors with encoders. A DC–DC buck-converter that steps down to 5 V is used to power a Raspberry Controller. The Raspberry Pi is connected to the two Roboclaws. The data from the Inertial Measurement Module(IMU) and wheel-odometry information from Roboclaw are used for dead-reckoning. Global localization of the robot is handled by the Marvel-minds beacon and GPS module associated with the robot. A Marvelmind indoor navigation beacon that is advantageously placed at the top center of the robot provides the necessary position information for GPS-denied navigation. A spring-return-based wheel suspension mechanism, designed according to the dimensions and kinematics of the robot, allows the robot to overcome high-rise obstacles and improves overall platform stability while navigating.

Figure 1.

Figure 1

The outline of the Raptor system of robots (a), the beacon platform (b), the payload platform (c), components associated with a beacon platform (d).

Control and Autonomy

The overall control architecture of the Raptor system is shown in Figure 2. The software nodes for the control and autonomy tasks are implemented over a Robot Operating System (ROS), running across a ROS master and integrates stationary beacons and other robot platforms. The ROS master controller is the primary controller that controls and overviews the rest of the sub-modules in the robot platform. A central workstation hosts the ROS master, and the beacon platforms run as slave devices. The slave devices are the ROS-enabled Raspberry Pi nodes that are in each robot platform. A user can establish a connection with the ROS master using a mobile or computer application over WiFi. The UWB (Ultra Wide Band) outdoor beacons by Marvel mind, an indoor positioning system, was used to find the robot location accurately up to 2 cm resolution. The positioning system works a minimum of 3 stationary beacons placed at different corners of the test-bed and a mobile beacon placed on top of the robot platform. Each beacon can either transmit or receive a signal. Therefore, Marvelmind has developed two localization architectures: Inverse Architecture (IA) and Non-Inverse Architecture (NIA). In IA, stationary beacons act as transmitters, and mobile beacons act as receivers with a maximum localization update rate of 16 Hz, halved with each additional mobile beacon added. In NIA, stationary beacons act as receivers, and mobile beacons act as transmitters, with the localization update rate remaining the same, albeit with a more complicated setup. This platform uses the IA approach for autonomous navigation. In addition to the UWB beacon, IMU and wheel odometry data are also used for the robot localization.

Figure 2.

Figure 2

The control architecture of the Raptor robot.

4. Path-Planning Strategy

The task of the Raptor robot is to inspect the maximum area in a false ceiling using a camera for rodent activities. Hence, the robot follows a vision-based complete coverage planning strategy (CCP). There are multiple approaches for achieving the CCP [38,39,40]. However, this study exploits the classic zig-zag path planning strategy to achieve maximum coverage for an area devoid of obstacles.

4.1. Robot Footprint Definition

Conventionally, the mobile robot’s footprint is the area swept by the robot when it is stationary. For an inspection robot, the footprint is defined as the effective area that can be inspected while stationary. In other terms, the footprint is the camera’s effective field of view. By definition, [37], the functional footprint is the optimal footprint associated with the robot that maximizes the task efficiency. Hence, the functional footprint (fr) can be considered as the optimal inspectable area using the usable field of view of the camera. On the other hand, the performance and efficiency of an inspection task can be weighed using the time taken for inspection, the area missed during the inspection, and the energy consumed by the robot during the inspection. Under the notion that traveling a minimal path for a given area will yield minimum time and minimal energy consumption, we considered area missed and path-length as the efficiency-determining factors in the scenario. To obtain the functional footprint of the robot, dependency on the functional footprint on every efficiency determining factor has been investigated. The parameters that define robot footprint are:

  1. Position of camera mounting in both the x and y-axis.

  2. The horizontal mounting angle α (yaw angle).

  3. The vertical mounting angle β (pitch angle).

Therefore, the fr (functional footprint) is represented by the optimal values for the camera’s position (x and y) and orientation (yaw angle—α and pitch angle—β) that results in the minimal area missed minimal total path length to be traveled during the inspection. Hence, the functional footprint can be represented as in Expression (1) given below:

fr=xyαβ (1)

The superscript ∗ on the variables represents that they hold optimal values.

4.2. Zig-Zag Based Area Coverage

Given a rectangular cross-section of size L×B, the whole region is divided into grid cells of size l×b. The size of the grid cells specified is split with respect to the camera’s field of view; the robot moves along each row, covering all the grid cells with the defined functional footprint of the camera. At the end of each row, the robot turns 90 degrees and moves to the next grid cell to reach the next row. This is repeated until the robot covers all the grid cells specified in the region up to the last row.

5. Optimal Coverage Strategy

This section details the derivation of the necessary objective functions for the minimum distance traveled while maximizing the area coverage for the false ceiling inspection using a zig-zag-based coverage strategy.

5.1. Total Zig-Zag Path Length Computation

Figure 3 shows the side-view of the robot from which we can geometrically extract the grid size for zig-zag path in terms of θ, β, and z. In Figure 3, z is the height at which the camera is positioned on the robot from the floor, β is the vertical angle of the camera, and θ represents the horizontal field of view of the camera. The length of the grid l can be calculated in terms of β and θ using Equation (2):

l=ztan(β+θ2); (2)

Figure 3.

Figure 3

The side-view of the robot with the field-of-view of the camera.

Figure 4 shows the top view of the robot with Xr and Yr as the x and y axes with respect to the robot frame. From Figure 4b, the breadth of the grid b can be geometrically derived. The b can be calculated in terms l and θ using Equation (3):

b=2l×tan(θ2) (3)

Figure 4.

Figure 4

The zig-zag area coverage strategy of the robot (a), area missed during the initial stage of operation (b), area missed at the turning (c).

For the zig-zag coverage strategy, the path length can be calculated in terms of distance covered in each row and distance covered during the transition between the rows, as shown in Figure 4a. Hence, the total path length of the robot can be represented using Equation (4):

Stotal=(Srows×nrows)+(Stransition×(nrows1)) (4)

where, Srows is the total distance covered along the rows, Stransition is the distance covered during the transition between the rows and is given by Equations (5) and (6):

Srows=(Lengthnrow(b2sin(α)+lcos(α)+y))nrows (5)

where y is the position of the camera in the y-axis and α, β are the orientation of the camera on the robot, Lengthnrow is the total length of each row strip, and nrows is the total number of row strips in the grid space. The term b2sin(α)+lcos(α) is the total horizontal distance covered by the robot with the change in orientation angle α of the camera.

Stransition=(2btan(θ/2))(nrows1) (6)

nrows is the total number of rows in the area of operation. The number of rows nrows can be calculated using Equation (7):

nrows=Bb (7)

The total distance (Stotal) covered by the robot for maximum area coverage can be written in terms of y, α and β using Equation (8):

Stotal(y,α,β)=(Lengthnrow(b2sin(α)+lcos(α)+y))nrows+(2btan(θ/2))(nrows1) (8)

From Equation (8), it is clear that the path length of the robot is dependent on parameters that define the footprint of the robot, which is α, β, and y. Hence, the functional footprint of the robot can be determined by estimating the optimal footprint parameters for the minimal path length for a given area of operation.

5.2. Total Area Missed

During the zig-zag path planning, loss in area coverage happens in two cases:

  1. The robot moves along the horizontal path while covering a row (Figure 4b).

  2. The robot does a transition between rows (Figure 4c).

The total area missed is formulated as the sum of the areas missed in case 1 and case 2. For case 1, the area missed is the initial area loss due to the robot’s starting position combined with area loss accumulated while the robot moves in the horizontal path. The initial area loss at the starting position can be estimated by calculating the areas represented by section 1, section 2, section 3, and section 4 represented in Figure 4b shown in Figure 4b. The area of section 1 can be estimated by calculating the area of the trapezoid as per Equation (9):

Areasection1=b2cos(α)+lsin(α)2×(b2sin(α)+lcos(α)) (9)

where b2cos(α) and lsin(α) are the two sides of the trapezoidal region and b2sin(α)+lcos(α) is the total horizontal length of section 1. Similarly, the area under section 2 can be calculated using the area of the rectangular region given in Equation (10):

Areasection2=((b2sin(α)+lcos(α))×x) (10)

The triangular area in section 3 is not considered in the estimation of initial area loss as most of the region in section 3 will be covered when the robot moves along the second row strip in a zig-zag path by overlapping a small region of the adjacent row strip. The accumulated area loss during the robot motion can be approximated with the area of section 4. The area of section 4 can be estimated using Equation (11)

Areasection4=(nrows×(Lengthnrow)×(lsin(α)+x)) (11)

where nrows are the total number of rows in the grid space, and Lengthnrow is the length of each row strip.

The area missed during the transition between rows can be approximated as the area of triangle ABC shown in Figure 4c. From Figure 4c, LineA meets with CircleO at point C, and LineB meets at the same CircleO at point B. Point A is the meeting point of LineB and LineA. The representation of LineA is given by Equation (12):

Y3×b2=0, (12)

and the representation of LineB is given by Equation (13):

X(l+y)=0 (13)

The equation for the CircleO can be represented as:

(xh)2+(yk)2=(l+y)2 (14)

where, h and k are the origins of the robots coordinates. The equation of the triangle ABC in terms of point A, B and C can be written as:

Areatriangle=12(x1(y2y3)+x2(y3y1)+x3(y1y2)) (15)

where, (x1,y1),(x2,y2) and (x3,y3) are the coordinates of points (A, B, C), which are obtained by solving a set of Equations (12)–(14). The total area missed, Equation (16), can be determined from Equations (9)–(11) and (15):

A(x,α,β)=(b2cos(α)+lsin(α)2)(b2sin(α)+lcos(α))+(b2sin(α)+lcos(α))x+nrows(Lengthnrow(lsin(α)+x))+((nrows1)(Areatriangle)) (16)

Similar to the equation for path-length (8), we can clearly observe that the area missed by the robot (16) is dependent on parameters that define the footprint of the robot, which are α, β, and x. Hence, the estimation of the functional footprint boils down to a multi-objective optimization problem with objectives given in Equation (17):

min(A(x,α,β))andmin(Stotal(y,α,β)) (17)

subject to the constraints:

xminxxmax;yminyymax;αminααmax;βminββmax. (18)

where xmin, xmax and ymin,ymax are the minimum and maximum possible position of camera on the robot; αmin,αmax and βmin,βmax are the minimum and maximum possible yaw angle and pitch angle to adjust the horizontal and vertical field of views. The parameters that shape the zig-zag path: (1) number of rows nrows (2) row length Srow and (3) transition length Stransition are computed from the optimal values for α, β, x and, y. The multi-objective optimization problem defined by Equations (17) and (18) is solved using a Genetic Algorithm (GA). GA is a subset of evolutionary algorithms that are widely used for solving complex multi-objective optimization problems. However, the genetic algorithms are modified to accommodate multiple objective functions for multi-objective optimization problems. There are different Multi-Objective Evolutionary algorithms (MOEA), including NSGA2 [41,42,43,44] and PESA-II [45,46]. In this work, we focused on NSGA-2 and MOEA/D algorithms for solving the constraint multi-objective optimization problem.

5.2.1. NSGA-2

The procedures involved in the NSGA2 algorithm are explained in the flow chart given in Figure 5. The NSGA2 algorithm yields a set of solution points, also known as the Pareto-Optimal solution set. The Pareto-front holds the dominant solutions for conflicting objectives functions. The NSGA2 algorithm is known for its simplicity of implementation and effectiveness. The NSGA2 algorithm is an improved version of NSGA using the fast non-dominated sorting concept [41].

Figure 5.

Figure 5

Procedure for the NSGA-2 algorithm implementation [41].

5.2.2. MOEA/D

The multi-objective evolutionary algorithm based on decomposition (MOEA/D) works by decomposing a given problem to multiple neighborhood sub-problems and finding the optimal solutions simultaneously [47,48]. MOEA/D is known to be more straightforward compared to other multi-objective optimization algorithms. Hence, we considered the usage of MOEA/D alongside NSGA2 for performing the concerned multi-objective optimization problem. The procedures involved in the MOEA/D algorithm are explained in the flow chart given in Figure 6.

Figure 6.

Figure 6

Procedure for MOEA/D algorithm implementation.

6. Results and Discussion

For the given scenario, we exploited genetic algorithms (GA) using NSGA2 (Non-Dominant Sorting Genetic Algorithm). We have conducted multiple simulations with different variations in population size of NSGA2 to solve for the functional footprint of the robot. In addition to NSGA2, a multi-objective evolutionary algorithm based on decomposition (MOEA/D) was also used to arrive at an optimal solution. The simulations were performed using the Pymoo library for NSGA2 and MOEA/D in Python3. The algorithms were executed on a computer with an Intel Core-i7 CPU and 16GB RAM. Part of the Section details the multiple simulation trials and experiments that were conducted.

6.1. NSGA2—Variation of Population

The NSGA2 optimization was completed for three different population sizes; 250, 500 and 2000. For the MOEA/D, the number neighbors and probability of neighbor mating are fixed as 15 and 0.7, respectively. For the area coverage, different area sizes of 2×2, 3×2 and 4×2 m in a false ceiling environment were considered for experimentation. In each set of simulations, the upper and lower bound of the constraints αmax, αmin, βmax, βmin, xmax, xmin, ymax, ymin were chosen as (0.0, 0.9547), (0.0, 0.687), (−0.15, 0.15) and (−0.10, 0.0) respectively. The simulation results for population 250 is given in Figure 7. The number of points belonging to the Pareto-front is less for a population of 250 compared to the other set of simulation trials. In the design space, the solutions were widely spread between (−0.145, 0.142), (−0.0941, −0.014), (0.0013, 0.499) and (0.019, 0.685) for x, y, α, and β. The simulation results for a population of 500 are given in Figure 8. The number of points corresponding to the Pareto-front is more for a population of 500 compared to a population of 250 in the previous simulation. In the design space, the solutions were spread across the intervals of (−0.1498, −0.00485), (−0.0939, −0.0067), (0.0002, 0.117) and (0.136, 0.6859) for x, y, α, and β. However, compared to simulation for a population of 250, the spread of solutions mostly converged to values −0.1498, −0.0939, 0.00 and 0.685 for x, y, α, and β. When it comes to the population of 2000, (given in Figure 9), the number of points belonging to the Pareto-front is more than that of simulations corresponding to populations of 250 and 500. In the design space, the solutions were less spread between (−0.1499, 0.0866), (−0.099, −0.0037), (0.00004, 0.365) and (0.124, 0.685) for x, y, α, and β. The values for x, y, α, and β are more converged to −0.1499, −0.0999, 0.003 and 0.685.

Figure 7.

Figure 7

NSGA2 optimization for area size 4 × 2 m with population size = 250.

Figure 8.

Figure 8

NSGA2 optimization for area size 4 × 2 m with population size = 500.

Figure 9.

Figure 9

NSGA2 optimization for area size 4 × 2 m with population size = 2000.

6.2. MOEA/D

The simulation results for the MOEA/D algorithm for a neighboring size of 15 are given in Figure 10. In the design space, the solutions were widely spread between (−0.1498, 0.0149), (−0.0931, −0.00), (0.0063, 0.499) and (0.0575, 0.685) for x, y, α, and β. Figure 11 shows the overall time taken to run each of the simulations by varying the population size and changing the optimization algorithm.

Figure 10.

Figure 10

MOEA/D optimization for area size 4 × 2 m with neighboring size 15.

Figure 11.

Figure 11

Time comparison for MOEA/D and variations of NSGA2 for finding the Pareto-front.

Table 1 summarizes the simulation results for varying population sizes in the optimization algorithm. The simulation for NSGA2-2000 takes at least 70% more time than the other optimization algorithm types. However, an optimization with a population size of 2000 for NSGA2 has shown better convergence results when compared with other population sizes and the MOEA/D algorithm. Therefore, we considered using NSGA2-2000 for further simulations by varying the area of the test-bed.

Table 1.

Consolidated simulation results for varying population size.

Simulation Results for Varying Population Size
Optimization Algorithm Time Taken (Seconds) Area Missed (Metres) Path Length (Metres)
NSGAii (population size—250) 0.0666 0.87 16.9
NSGAii (population size—500) 0.182 0.332 15.2
NSGAii (population size—2000) 1.424 0.16 14.6
MOEA/D (neighboring size—15) 0.414 0.37 15.7

6.3. NSGA-2 Variation of Area

The NSGA2-2000 optimization was completed for three different test-bed sizes of 2 × 2, 3 × 2 and 4 × 2 m. For each of the simulation results, total area missed and total path length for the robot to maximize area coverage were computed. Figure 12, Figure 13 and Figure 14 show the Pareto-chart using NSGA2-2000 for each of the area sizes. The solutions for all of the test-bed sizes were less spread between (−0.1499, 0.006), (−0.0999, −0.0005), (0.0002, 0.378) and (0.023, 0.6859) for x, y, α, and β. The values for x, y, α, and β are more converged to −0.1499, −0.0999, 0.0002 and 0.685.

Figure 12.

Figure 12

NSGA2 optimization for area size 2 × 2 m with population size = 2000.

Figure 13.

Figure 13

NSGA2 optimization for area size 3 × 2 m with population size = 2000.

Figure 14.

Figure 14

NSGA2 optimization for area size 4 × 2 m with population size = 2000.

Table 2 summarizes the simulation results with overall area coverage performance and path length of the robot to maximize area coverage using NSGA2-2000 for different test-bed sizes.

Table 2.

Consolidated simulation results for varying area.

Simulation Results for Varying Area Using NSGA2-2000
Test-Bed Size (Metres) Srow (Metres) Stransition (Metres) Area Missed (Metres) Path Length (Metres)
2 × 2 1.401 0.4951 0.19 7.089
3 × 2 2.401 0.4951 0.21 11.08
4 × 2 3.401 0.4951 0.16 15.08

6.4. Experiments

The experiments were carried out on the Raptor robot to validate the simulation results for the optimal placement of the camera to maximize area coverage and minimize path length. The robot was operated inside a false ceiling test-bed of a total area of 4×2 m. Multiple experiments were performed based by reducing the area of operation to 2×2 (Figure 15), 3×2 (Figure 16) and 4×2 m (Figure 17) within the same test-bed. We installed MarvelMind beacons on different locations of the test-bed for robot localization. The total path length for each trial is recorded from the experiment results. We estimated the robot pose by fusing the beacon position, wheel odometry, and inertial measurements from the IMU sensor. The row length, transition length between the rows obtained after performing the optimization are programmed on the robot for performing the zig-zag.

Figure 15.

Figure 15

The experiment trial performed in the false-ceiling test-bed for covering 2×2 m. (a) Robot performing inspection in the false-ceiling test-bed at various time frames. (b) Fused odometry plot of robot during experimental trial.

Figure 16.

Figure 16

The experiment trial performed in the false-ceiling test-bed for covering 3×2 m. (a) Robot performing inspection in the false-ceiling test-bed at various time frames. (b) Fused odometry plot of robot during experimental trial.

Figure 17.

Figure 17

The experiment trial performed in the false-ceiling test-bed for covering 4×2 m. (a) Robot performing inspection in the false-ceiling test-bed at various time frames. (b) Fused odometry plot of the robot during the experimental trial.

Table 3 summarizes the experimental data obtained. Time taken for the robot to cover 2×2, 3×2, and 4×2 m are 61.132, 80.177, and 120.43 s, respectively. Correspondingly, path-lengths for 2×2, 3×2, and 4×2 m were 8.042, 12.542, and 15.662 m, respectively. The observed results agree with the simulation results for 2, 3×2, and 4×2 m.

Table 3.

Consolidated experimental results for varying areas.

Experimental Results for Varying Area Using Solutions from NSGA2-2000
Test-Bed Size (Metres) Time Taken (Seconds) Path Length (Metres)
2 × 2 68.37 7.452
3 × 2 82.237 11.97
4 × 2 117.28 15.96

7. Conclusions and Future Works

This work calculates an optimal footprint for an in-house developed inspection robot called Raptor, where the footprint is defined by the optimal position and orientation of a camera on the robot. We considered the minimization of area-missed during the inspection task and path length of the robot for different positions and orientations of the camera. The mathematical expression for area missed and path covered for a zig-zag area-coverage strategy was derived, and optimal values for camera position coordinates and orientation angles were determined using constraint multi-objective optimization. We utilized multi-objective optimization tools, such NSGA-2 and MOEA/D, for performing multi-objective optimization. The results of multi-objective optimization were validated by conducting experiments on a false-ceiling test-bed using the Raptor robot. The future works of this research will focus on the dynamic optimization, consideration of energy and multiple path-planning strategies for defining functional footprint-based path generation considering static and dynamic obstacles.

Author Contributions

Conceptualization, T.P.; methodology, T.P. and V.S.; software, V.S. and T.P.; validation, S.G.A.P., M.R.E., and T.T.T.; analysis, V.S., S.G.A.P., and M.R.E.; original draft, T.P., V.S., and S.G.A.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the National Robotics Programme under its Robotics Enabling Capabilities and Technologies (Funding Agency Project No. 1922500051), the National Robotics Programme under its Robot Domain Specific (Funding Agency Project No. 192 22 00058), the National Robotics Programme under its Robotics Domain Specific (Funding Agency Project No. 1922200108), and administered by the Agency for Science, Technology and Research.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data sharing not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Footnotes

Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

References

  • 1.Pest Control Market by Pest Type. [(accessed on 2 June 2021)]; Available online: https://www.marketsandmarkets.com/Market-Reports/pest-control-market-144665518.html.
  • 2.QUINCE Market Insights. [(accessed on 2 June 2021)]; Available online: https://www.globenewswire.com/news-release/2021/05/24/2234537/0/en/Global-Pest-Control-Market-is-Estimated-to-Grow-at-a-CAGR-of-5-25-from-2021-to-2030.html.
  • 3.Rabiee M.H., Mahmoudi A., Siahsarvie R., Kryštufek B., Mostafavi E. Rodent-borne diseases and their public health importance in Iran. PLoS Negl. Trop. Dis. 2018;12:e0006256. doi: 10.1371/journal.pntd.0006256. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Nimo-Paintsil S.C., Fichet-Calvet E., Borremans B., Letizia A.G., Mohareb E., Bonney J.H., Obiri-Danso K., Ampofo W.K., Schoepp R.J., Kronmann K.C. Rodent-borne infections in rural Ghanaian farming communities. PLoS ONE. 2019;14:e0215224. doi: 10.1371/journal.pone.0215224. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Meerburg B.G., Singleton G.R., Kijlstra A. Rodent-borne diseases and their risks for public health. Crit. Rev. Microbiol. 2009;35:221–270. doi: 10.1080/10408410902989837. [DOI] [PubMed] [Google Scholar]
  • 6.Costa F., Ribeiro G.S., Felzemburgh R.D., Santos N., Reis R.B., Santos A.C., Fraga D.B.M., Araujo W.N., Santana C., Childs J.E., et al. Influence of household rat infestation on Leptospira transmission in the urban slum environment. PLoS Negl. Trop. Dis. 2014;8:e3338. doi: 10.1371/journal.pntd.0003338. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Tapsoba C. Rodents Damage Cable, Causing Internet Failure. [(accessed on 21 June 2016)]; Available online: http://www.northwesttalon.org/2017/03/03/rodents-damage-cable-causing-internet-failure/
  • 8.Almeida A., Corrigan R., Sarno R. The economic impact of commensal rodents on small businesses in Manhattan’s Chinatown: Trends and possible causes. Suburb. Sustain. 2013;1:2. doi: 10.5038/2164-0866.1.1.2. [DOI] [Google Scholar]
  • 9.Jassat W., Naicker N., Naidoo S., Mathee A. Rodent control in urban communities in Johannesburg, South Africa: From research to action. Int. J. Environ. Health Res. 2013;23:474–483. doi: 10.1080/09603123.2012.755156. [DOI] [PubMed] [Google Scholar]
  • 10.Kariduraganavar R.M., Yadav S.S. Emerging Trends in Photonics, Signal Processing and Communication Engineering. Springer; Berlin/Heidelberg, Germany: 2020. Design and Implementation of IoT Based Rodent Monitoring and Avoidance System in Agricultural Storage Bins; pp. 233–241. [Google Scholar]
  • 11.Ross R., Parsons L., Thai B.S., Hall R., Kaushik M. An IOT smart rodent bait station system utilizing computer vision. Sensors. 2020;20:4670. doi: 10.3390/s20174670. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Parsons L., Ross R. A computer-vision approach to bait level estimation in rodent bait stations. Comput. Electron. Agric. 2020;172:105340. doi: 10.1016/j.compag.2020.105340. [DOI] [Google Scholar]
  • 13.Sowmika T., Malathi G. IOT Based Smart Rodent Detection and Fire Alert System in Farmland. Int. Res. J. Multidiscip. Technovation. 2020:1–6. doi: 10.34256/irjmt2031. [DOI] [Google Scholar]
  • 14.Cambra C., Sendra S., Garcia L., Lloret J. Low cost wireless sensor network for rodents detection; Proceedings of the 2017 10th IFIP Wireless and Mobile Networking Conference (WMNC); Valencia, Spain. 25–27 September 2017; pp. 1–7. [Google Scholar]
  • 15.Walsh J.R., Lynch P.J. Remote Sensing Rodent Bait Station Tray. 10,897,887. U.S. Patent. 2021 Jan 26;
  • 16.Barnett S. Rodent Pest Management. CRC Press; Boca Raton, FL, USA: 2018. Exploring, sampling, neophobia, and feeding; pp. 295–320. [Google Scholar]
  • 17.Feng A.Y., Himsworth C.G. The secret life of the city rat: A review of the ecology of urban Norway and black rats (Rattus norvegicus and Rattus rattus) Urban Ecosyst. 2014;17:149–162. doi: 10.1007/s11252-013-0305-4. [DOI] [Google Scholar]
  • 18.Ma Q., Tian G., Zeng Y., Li R., Song H., Wang Z., Gao B., Zeng K. Pipeline In-Line Inspection Method, Instrumentation and Data Management. Sensors. 2021;21:3862. doi: 10.3390/s21113862. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Bahnsen C.H., Johansen A.S., Philipsen M.P., Henriksen J.W., Nasrollahi K., Moeslund T.B. 3D Sensors for Sewer Inspection: A Quantitative Review and Analysis. Sensors. 2021;21:2553. doi: 10.3390/s21072553. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Vidal V.F., Honório L.M., Dias F.M., Pinto M.F., Carvalho A.L., Marcato A.L.M. Sensors Fusion and Multidimensional Point Cloud Analysis for Electrical Power System Inspection. Sensors. 2020;20:4042. doi: 10.3390/s20144042. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Yin S., Ren Y., Zhu J., Yang S., Ye S. A Vision-Based Self-Calibration Method for Robotic Visual Inspection Systems. Sensors. 2013;13:16565–16582. doi: 10.3390/s131216565. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Venkateswaran S., Chablat D., Hamon P. An optimal design of a flexible piping inspection robot. J. Mech. Robot. 2021;13:035002. doi: 10.1115/1.4049948. [DOI] [Google Scholar]
  • 23.Imran M.H., Zim M.Z.H., Ahmmed M. Proceedings of International Joint Conference on Advances in Computational Intelligence. Springer; Berlin/Heidelberg, Germany: 2021. PIRATE: Design and Implementation of Pipe Inspection Robot; pp. 77–88. [Google Scholar]
  • 24.Zhonglin Z., Bin F., Liquan L., Encheng Y. Design and Function Realization of Nuclear Power Inspection Robot System. Robotica. 2021;39:165–180. doi: 10.1017/S0263574720000740. [DOI] [Google Scholar]
  • 25.Parween R., Muthugala M., Heredia M.V., Elangovan K., Elara M.R. Collision Avoidance and Stability Study of a Self-Reconfigurable Drainage Robot. Sensors. 2021;21:3744. doi: 10.3390/s21113744. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Tu X., Xu C., Liu S., Lin S., Chen L., Xie G., Li R. LiDAR Point Cloud Recognition and Visualization with Deep Learning for Overhead Contact Inspection. Sensors. 2020;20:6387. doi: 10.3390/s20216387. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Ajeil F.H., Ibraheem I.K., Azar A.T., Humaidi A.J. Grid-based mobile robot path planning using aging-based ant colony optimization algorithm in static and dynamic environments. Sensors. 2020;20:1880. doi: 10.3390/s20071880. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Delmerico J., Mueggler E., Nitsch J., Scaramuzza D. Active autonomous aerial exploration for ground robot path planning. IEEE Robot. Autom. Lett. 2017;2:664–671. doi: 10.1109/LRA.2017.2651163. [DOI] [Google Scholar]
  • 29.Araki B., Strang J., Pohorecky S., Qiu C., Naegeli T., Rus D. Multi-robot path planning for a swarm of robots that can both fly and drive; Proceedings of the 2017 IEEE International Conference on Robotics and Automation (ICRA); Singapore. 29 May–3 June 2017; pp. 5575–5582. [Google Scholar]
  • 30.Krajník T., Majer F., Halodová L., Vintr T. Navigation without localisation: Reliable teach and repeat based on the convergence theorem; Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS); Madrid, Spain. 1–5 October 2018; pp. 1657–1664. [Google Scholar]
  • 31.Markom M., Adom A., Shukor S.A., Rahim N.A., Tan E.M.M., Irawan A. Scan matching and KNN classification for mobile robot localisation algorithm; Proceedings of the 2017 IEEE 3rd International Symposium in Robotics and Manufacturing Automation (ROMA); Kuala Lumpur, Malaysia. 19–21 September 2017; pp. 1–6. [Google Scholar]
  • 32.Watiasih R., Rivai M., Wibowo R.A., Penangsang O. Path planning mobile robot using waypoint for gas level mapping; Proceedings of the 2017 International Seminar on Intelligent Technology and Its Applications (ISITIA); Surabaya, Indonesia. 28–29 August 2017; pp. 244–249. [Google Scholar]
  • 33.Miao X., Lee J., Kang B.Y. Scalable coverage path planning for cleaning robots using rectangular map decomposition on large environments. IEEE Access. 2018;6:38200–38215. doi: 10.1109/ACCESS.2018.2853146. [DOI] [Google Scholar]
  • 34.Sasaki T., Enriquez G., Miwa T., Hashimoto S. Adaptive path planning for cleaning robots considering dust distribution. J. Robot. Mechatron. 2018;30:5–14. doi: 10.20965/jrm.2018.p0005. [DOI] [Google Scholar]
  • 35.Majeed A., Lee S. A new coverage flight path planning algorithm based on footprint sweep fitting for unmanned aerial vehicle navigation in urban environments. Appl. Sci. 2019;9:1470. doi: 10.3390/app9071470. [DOI] [Google Scholar]
  • 36.Yu K., Guo C., Yi J. Complete and Near-Optimal Path Planning for Simultaneous Sensor-Based Inspection and Footprint Coverage in Robotic Crack Filling; Proceedings of the 2019 International Conference on Robotics and Automation (ICRA); Montreal, QC, Canada. 20–24 May 2019; pp. 8812–8818. [Google Scholar]
  • 37.Pathmakumar T., Rayguru M.M., Ghanta S., Kalimuthu M., Elara M.R. An Optimal Footprint Based Coverage Planning for Hydro Blasting Robots. Sensors. 2021;21:1194. doi: 10.3390/s21041194. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38.Choset H. Coverage of known spaces: The boustrophedon cellular decomposition. Auton. Robot. 2000;9:247–253. doi: 10.1023/A:1008958800904. [DOI] [Google Scholar]
  • 39.Choi Y.H., Lee T.K., Baek S.H., Oh S.Y. Online complete coverage path planning for mobile robots based on linked spiral paths using constrained inverse distance transform; Proceedings of the 2009 IEEE/RSJ International Conference on Intelligent Robots and Systems; St. Louis, MO, USA. 10–15 October 2009; pp. 5788–5793. [Google Scholar]
  • 40.Cabreira T.M., Di Franco C., Ferreira P.R., Buttazzo G.C. Energy-aware spiral coverage path planning for uav photogrammetric applications. IEEE Robot. Autom. Lett. 2018;3:3662–3668. doi: 10.1109/LRA.2018.2854967. [DOI] [Google Scholar]
  • 41.Deb K., Pratap A., Agarwal S., Meyarivan T. A fast and elitist multiobjective genetic algorithm: NSGA-II. IEEE Trans. Evol. Comput. 2002;6:182–197. doi: 10.1109/4235.996017. [DOI] [Google Scholar]
  • 42.Cai X., Wang P., Du L., Cui Z., Zhang W., Chen J. Multi-objective three-dimensional DV-hop localization algorithm with NSGA-II. IEEE Sens. J. 2019;19:10003–10015. doi: 10.1109/JSEN.2019.2927733. [DOI] [Google Scholar]
  • 43.Aminmahalati A., Fazlali A., Safikhani H. Multi-objective optimization of CO boiler combustion chamber in the RFCC unit using NSGA II algorithm. Energy. 2021;221:119859. doi: 10.1016/j.energy.2021.119859. [DOI] [Google Scholar]
  • 44.Ghaderian M., Veysi F. Multi-objective optimization of energy efficiency and thermal comfort in an existing office building using NSGA-II with fitness approximation: A case study. J. Build. Eng. 2021;41:102440. doi: 10.1016/j.jobe.2021.102440. [DOI] [Google Scholar]
  • 45.Corne D.W., Jerram N.R., Knowles J.D., Oates M.J. PESA-II: Region-based selection in evolutionary multiobjective optimization; Proceedings of the 3rd Annual Conference on Genetic and Evolutionary Computation; San Francisco, CA, USA. 7–11 July 2001. [Google Scholar]
  • 46.Ridha H.M., Gomes C., Hizam H., Ahmadipour M., Muhsen D.H., Ethaib S. Optimum design of a standalone solar photovoltaic system based on novel integration of iterative-PESA-II and AHP-VIKOR methods. Processes. 2020;8:367. doi: 10.3390/pr8030367. [DOI] [Google Scholar]
  • 47.Zhang Q., Li H. MOEA/D: A multiobjective evolutionary algorithm based on decomposition. IEEE Trans. Evol. Comput. 2007;11:712–731. doi: 10.1109/TEVC.2007.892759. [DOI] [Google Scholar]
  • 48.Wang G., Li X., Gao L., Li P. Energy-efficient distributed heterogeneous welding flow shop scheduling problem using a modified MOEA/D. Swarm Evol. Comput. 2021;62:100858. doi: 10.1016/j.swevo.2021.100858. [DOI] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

Data sharing not applicable.


Articles from Sensors (Basel, Switzerland) are provided here courtesy of Multidisciplinary Digital Publishing Institute (MDPI)

RESOURCES