Abstract
This paper presents an innovative adaptive control system for collaborative sorting robotic arms achieving three key technical breakthroughs: (1) a multimodal sensor fusion algorithm integrating vision, force, and position sensors with dynamic reliability weighting achieving 98.7% sorting accuracy, (2) a distributed edge computing architecture enabling 3.2ms average response time through local processing optimization, and (3) adaptive control mechanisms with online learning capabilities delivering 847 items/hour throughput capacity. The system combines advanced fusion algorithms with machine learning techniques to optimize performance under varying operational conditions including payload changes, environmental disturbances, and collaborative coordination requirements. Experimental validation demonstrates 15% improvement in accuracy over commercial systems, 60% reduction in communication latency compared to centralized architectures, and robust operation with maintained performance above 85% during sensor failures. The research contributes to intelligent manufacturing advancement by providing a scalable framework for collaborative robotic sorting applications that adapts to dynamic production requirements while maintaining optimal performance metrics in Industry 4.0 implementations.
Keywords: Collaborative robotics, Multimodal sensor fusion, Edge computing, Adaptive control, Industrial automation, Intelligent sorting
Subject terms: Engineering, Mathematics and computing
Introduction
With the rapid advancement of Industry 4.0 and intelligent manufacturing paradigms, automated production line systems have become fundamental pillars for enhancing manufacturing efficiency and product quality1. The integration of robotic manipulators in sorting operations represents a critical component of smart factory ecosystems, where precision, speed, and adaptability determine competitive advantages in modern manufacturing environments2. As global supply chains demand increasingly flexible and responsive production capabilities, the development of sophisticated robotic sorting systems has emerged as a pivotal research domain that bridges theoretical control algorithms with practical industrial applications3.
Current robotic arm sorting systems face significant technical challenges that limit their effectiveness in dynamic production environments. Traditional vision-based sorting approaches suffer from limitations in complex lighting conditions, object occlusion scenarios, and multi-target recognition tasks that require real-time decision-making capabilities4.
Recent advances in visual servoing with field-of-view constraints have demonstrated effective force-position control integration for single manipulators operating without complete structural knowledge5. These methods address visual servoing under constrained viewing conditions but focus primarily on individual robotic arms. The collaborative sorting environment introduces additional complexity requiring multimodal sensor fusion capabilities that compensate for field-of-view limitations through force and position sensor integration, enabling robust operation when visual information is partially occluded by neighboring robots or workspace obstacles.
Furthermore, conventional control architectures often struggle with multimodal sensor integration, where the fusion of visual, tactile, and proximity sensing data remains computationally intensive and prone to latency issues that compromise sorting accuracy and throughput6. Existing collaborative robotics systems in manufacturing environments typically rely on centralized processing architectures that introduce communication bottlenecks and limit real-time responsiveness7. Recent developments in edge computing for industrial IoT applications have shown promise for distributed processing, but their integration with adaptive control systems for collaborative manipulation tasks remains largely unexplored8.
The lack of adaptive control mechanisms in existing systems also presents substantial barriers to handling variations in object properties, environmental conditions, and collaborative workspace requirements where multiple robotic units must coordinate their operations seamlessly9.
Edge computing technologies offer promising solutions to address the computational and latency challenges inherent in intelligent sorting systems by enabling distributed processing capabilities that reduce communication overhead and improve real-time performance10. The integration of multimodal sensor fusion algorithms with edge computing architectures presents opportunities to develop adaptive control systems that can process heterogeneous sensor data locally while maintaining high-frequency control loops essential for precise manipulation tasks11. This technological convergence enables the development of collaborative robotic sorting systems that can adapt to changing production requirements while maintaining optimal performance metrics across diverse operational scenarios.
The primary objective of this research is to address three critical technical challenges in industrial collaborative sorting: (1) insufficient sensor fusion accuracy under varying environmental conditions (current systems achieve < 85% accuracy), (2) excessive communication latency in centralized architectures (> 15ms response times), and (3) limited adaptability to dynamic production requirements. This research develops a comprehensive adaptive control system that achieves 98.7% sorting accuracy, 3.2ms response time, and maintains > 85% performance during sensor failures through three key technical innovations: dynamic multimodal sensor fusion with adaptive reliability weighting, distributed edge computing architecture optimized for real-time control, and adaptive control mechanisms with online learning capabilities.
The proposed system aims to integrate computer vision, force sensing, and proximity detection capabilities within a distributed edge computing framework that enables real-time sensor data fusion and adaptive control parameter adjustment. Key research contributions include the development of novel sensor fusion algorithms optimized for edge computing platforms, the design of adaptive control strategies that respond to dynamic sorting scenarios, and the implementation of collaborative coordination protocols that enable multiple robotic units to operate efficiently within shared workspaces.
The technical approach encompasses three primary research directions with specific innovations beyond existing collaborative robotics approaches: first, the development of multimodal sensor fusion algorithms that combine visual, tactile, and proximity sensing modalities with dynamic reliability weighting to achieve comprehensive object recognition and pose estimation capabilities; second, the design of distributed edge computing architectures that support real-time sensor data processing while maintaining sub-5ms control loop execution through optimized resource allocation; and third, the implementation of adaptive control strategies with online learning capabilities that adjust system parameters based on environmental feedback and performance metrics to optimize sorting operations under varying conditions.
The main contributions of this research are: (1) Novel multimodal sensor fusion algorithms optimized for edge computing platforms with adaptive weighting mechanisms that achieve 15% improvement in accuracy over single-modality approaches; (2) Distributed edge computing architecture enabling deterministic real-time performance with 60% reduction in communication latency compared to centralized systems; (3) Adaptive control strategies incorporating machine learning techniques for continuous performance optimization with stability guarantees; (4) Comprehensive experimental validation demonstrating superior performance compared to three commercial sorting systems in realistic production environments.
This paper is structured to present a systematic investigation of the proposed adaptive control system across six main sections. Following this introduction, Section II provides a comprehensive literature review of existing robotic sorting technologies and identifies research gaps in current approaches. Section III details the system architecture and methodology, including sensor fusion algorithms and edge computing implementation strategies. Section IV presents experimental validation results demonstrating system performance under various operational scenarios. Section V discusses the implications of findings and potential applications in industrial settings. Finally, Section VI concludes with a summary of contributions and directions for future research development.
Theoretical foundation of multimodal sensor fusion and edge computing
Multimodal sensor data fusion algorithms
Multimodal sensor systems in robotic sorting applications integrate heterogeneous data streams from vision sensors, force sensors, and position sensors to create comprehensive environmental perception capabilities that exceed the limitations of individual sensing modalities12. Vision sensors provide high-dimensional spatial information including object geometry, surface features, and relative positioning data that enable precise target identification and pose estimation in complex sorting scenarios. Force sensors contribute tactile feedback through multi-axis force and torque measurements that facilitate contact detection, grip force optimization, and object manipulation safety monitoring during sorting operations13. Position sensors deliver real-time kinematic data including joint angles, end-effector coordinates, and velocity profiles that support accurate trajectory planning and collision avoidance in collaborative workspace environments.
The integration of multimodal sensor data requires sophisticated fusion algorithms that can handle temporal synchronization, measurement uncertainties, and computational constraints inherent in real-time robotic control systems14.
The fusion implementation utilizes a modified weighted Bayesian framework with adaptive reliability assessment. The dynamic reliability scores for each sensor modality are computed as:
![]() |
1 |
where
represents the reliability score for sensor
at time
,
denotes accuracy metric,
indicates normalized noise level,
represents environmental suitability factor, and
,
,
are weighting coefficients. The reliability scores are dynamically updated using exponential moving averages:
![]() |
2 |
where
provides optimal balance between responsiveness and stability based on experimental validation.
The proposed fusion architecture employs a three-stage processing pipeline as illustrated in Fig. 1: (1) sensor data preprocessing and temporal alignment, (2) reliability assessment and dynamic weighting computation, and (3) probabilistic fusion with uncertainty quantification. This approach enables robust operation when individual sensors experience degraded performance due to environmental conditions or partial failures.
Fig. 1.
Detailed sensor fusion algorithm flowchart showing the three-stage processing pipeline with decision points for adaptive weighting and fault tolerance mechanisms.
Kalman filtering represents a fundamental approach for linear sensor fusion applications where system dynamics and measurement models follow Gaussian distributions. The standard Kalman filter algorithm incorporates prediction and correction phases to estimate system states based on prior knowledge and current sensor measurements according to the following state estimation equation:
![]() |
3 |
where
represents the posterior state estimate,
denotes the Kalman gain matrix,
indicates the measurement vector, and
represents the observation model matrix15.
Particle filtering algorithms address nonlinear sensor fusion challenges by representing probability distributions through weighted particle sets that approximate complex posterior distributions without Gaussian assumptions. The particle filter recursively updates particle weights based on likelihood functions that incorporate multimodal sensor observations, enabling robust state estimation in scenarios with nonlinear dynamics and non-Gaussian noise characteristics. The importance sampling approach in particle filtering follows the sequential Monte Carlo framework where particle weights are updated according to:
![]() |
4 |
where
represents the weight of particle
at time
,
denotes the likelihood function, and
indicates the proposal distribution16.
Bayesian fusion frameworks provide theoretical foundations for combining multiple sensor modalities through probabilistic inference mechanisms that quantify measurement uncertainties and enable optimal decision-making under incomplete information conditions. The Bayesian approach incorporates prior knowledge about system states and sensor characteristics to compute posterior probability distributions that reflect the most likely system configurations given available sensor data. Multi-sensor Bayesian fusion employs conditional independence assumptions to simplify computational complexity while maintaining estimation accuracy through the following joint probability formulation:
![]() |
5 |
where
represents the posterior distribution given multiple sensor observations,
denotes the prior distribution, and
indicates individual sensor likelihood functions.
The mathematical modeling of multimodal sensor fusion systems requires consideration of temporal alignment, coordinate transformations, and measurement noise characteristics that influence fusion performance and computational requirements17. Sensor data preprocessing involves calibration procedures that establish geometric relationships between different sensor coordinate frames and temporal synchronization algorithms that align asynchronous sensor measurements to common time references. The fusion strategy selection depends on application-specific requirements including accuracy specifications, computational constraints, and real-time performance targets that determine the optimal balance between estimation quality and processing efficiency. Advanced fusion architectures incorporate adaptive weighting mechanisms that adjust sensor contributions based on dynamic reliability assessments and environmental conditions to maintain robust performance across diverse operational scenarios.
Edge computing architecture and real-time processing technology
Edge computing architectures in Industrial Internet of Things (IIoT) environments provide significant advantages over centralized cloud computing approaches by reducing communication latency, enhancing data privacy, and enabling autonomous operation capabilities that are critical for real-time robotic control applications18. The proximity of edge computing nodes to sensor sources and actuator systems minimizes network transmission delays that can compromise control loop stability and sorting precision in time-sensitive manufacturing operations. Edge computing infrastructures also facilitate distributed intelligence deployment where computational resources are strategically positioned throughout production facilities to support local decision-making processes without dependence on external network connectivity or remote server availability19.
Computational resource allocation strategies in edge computing environments require sophisticated optimization algorithms that balance processing demands across heterogeneous edge nodes while maintaining quality of service requirements for real-time control applications. The resource allocation problem can be formulated as a multi-objective optimization challenge where computing capacity, memory utilization, and energy consumption constraints must be simultaneously satisfied.
The edge computing optimization problem is formulated with the following variables and constraints:
Optimization Variables:
: binary allocation variable for task
on node
-
: resource utilization rate for node
at time t -
: bandwidth allocation for task
on node 
Constraints: Task assignment:
- Capacity limits:
- Deadline constraints: 
The optimization objective function incorporates task completion time, resource utilization efficiency, and system reliability:
![]() |
6 |
where
represents task completion time,
denotes resource utilization cost,
indicates energy consumption,
represents binary allocation variables, and
,
,
are weighting coefficients for different optimization objectives20.
The relationship between optimization and control is established through real-time resource allocation that directly affects control loop execution times and system responsiveness.
Task scheduling strategies for edge computing nodes employ dynamic prioritization mechanisms that consider real-time constraints, task dependencies, and resource availability to optimize overall system performance in robotic sorting applications. Priority-based scheduling algorithms assign execution priorities to incoming sensor data processing tasks, control algorithm computations, and inter-node communication operations based on their criticality to sorting accuracy and safety requirements. The task scheduling framework incorporates deadline-aware algorithms that guarantee timely completion of critical control functions while efficiently utilizing available computational resources for non-critical background processes.
The edge computing framework designed for robotic arm control systems integrates distributed processing capabilities with hierarchical control architectures that enable seamless coordination between local edge nodes and centralized supervisory systems. Local edge nodes execute high-frequency control loops, sensor data preprocessing, and immediate response algorithms that require minimal latency, while higher-level coordination tasks and optimization algorithms are distributed across multiple edge nodes or delegated to cloud resources when computational demands exceed local capabilities. The framework employs containerized microservices architectures that facilitate scalable deployment and dynamic load balancing across edge computing infrastructure.
Real-time data processing mechanisms in edge computing environments utilize stream processing algorithms and event-driven architectures to handle continuous sensor data flows with deterministic latency guarantees21. The data processing pipeline incorporates buffering strategies, parallel processing techniques, and adaptive sampling methods that maintain consistent performance under varying computational loads and network conditions. Real-time processing requirements are satisfied through predictable execution scheduling where critical control tasks receive guaranteed processor time allocation according to the following real-time constraint:
![]() |
7 |
where
represents task deadline,
denotes worst-case execution time,
indicates maximum jitter, and
represents blocking time from lower-priority tasks.
The edge computing architecture incorporates fault tolerance mechanisms and redundancy strategies that ensure continuous operation despite individual node failures or network disruptions that could compromise robotic sorting operations22. Distributed consensus algorithms enable coordinated decision-making across multiple edge nodes while maintaining consistency in shared state information and control commands. The fault tolerance framework employs checkpointing mechanisms, state replication, and graceful degradation strategies that preserve essential sorting functionality even under partial system failures.
Communication protocols between edge nodes utilize lightweight messaging frameworks optimized for low-latency data exchange and efficient bandwidth utilization in industrial network environments. The inter-node communication architecture supports both synchronous and asynchronous message passing paradigms that accommodate different types of coordination requirements ranging from immediate control responses to periodic status updates. Quality of service mechanisms ensure that critical control messages receive priority treatment over routine data transmissions, maintaining reliable communication channels for safety–critical robotic operations. The communication latency optimization follows the constraint:
![]() |
8 |
where
represents total communication latency,
denotes processing delay,
indicates transmission time,
represents queueing delay, and
defines the maximum acceptable latency for real-time control applications.
Robotic arm adaptive control theory
Robotic arm kinematics modeling establishes the mathematical relationships between joint configurations and end-effector positions that form the foundation for precise motion planning and control in collaborative sorting applications23. Forward kinematics analysis utilizes Denavit-Hartenberg parameter conventions to derive transformation matrices that describe the spatial relationships between consecutive joint coordinate frames, enabling accurate computation of end-effector pose based on joint angle measurements. Inverse kinematics solutions provide the mathematical framework for determining required joint configurations to achieve desired end-effector positions and orientations, which is essential for trajectory planning algorithms that guide robotic arms through complex sorting motions while avoiding workspace obstacles and maintaining operational constraints24.
Dynamic modeling of robotic manipulators incorporates the effects of inertial forces, gravitational loads, Coriolis forces, and friction phenomena that influence system behavior during high-speed sorting operations. The Lagrangian formulation provides a systematic approach for deriving equations of motion that capture the complex nonlinear dynamics inherent in multi-degree-of-freedom robotic systems. The general form of robotic arm dynamics follows the equation:
![]() |
9 |
where
represents the inertia matrix,
denotes the Coriolis and centripetal force matrix,
indicates gravitational forces,
represents friction forces,
denotes joint positions, and
represents applied joint torques15.
Adaptive control algorithms address parameter uncertainties and time-varying characteristics in robotic systems by continuously adjusting controller parameters based on system performance feedback and estimation errors. Model Reference Adaptive Control (MRAC) techniques employ reference models that define desired system behavior and adaptation mechanisms that modify controller gains to minimize tracking errors between actual and reference system responses. The adaptive control framework incorporates Lyapunov stability analysis to guarantee convergence and bounded-input bounded-output stability even in the presence of modeling uncertainties and external disturbances that commonly occur in industrial sorting environments14.
Fuzzy logic control systems provide robust solutions for handling nonlinear dynamics and uncertain parameters in robotic arm control applications through linguistic rule-based reasoning mechanisms that mimic human expert knowledge. Fuzzy controllers utilize membership functions and inference engines to map sensor inputs and control objectives into appropriate actuator commands without requiring precise mathematical models of system dynamics. The fuzzy control approach demonstrates particular effectiveness in collaborative sorting scenarios where environmental conditions, object properties, and task requirements exhibit significant variability that challenges traditional model-based control strategies25.
Neural network control architectures leverage learning capabilities and nonlinear approximation properties to develop adaptive control systems that can handle complex robotic arm dynamics without explicit mathematical modeling. Multilayer perceptron networks and radial basis function networks serve as universal function approximators that can learn inverse dynamics models through supervised training processes using collected input–output data from robotic system operations. Neural adaptive control combines neural network learning with stability-guaranteed adaptation laws to ensure stable performance during online learning phases while maintaining trajectory tracking accuracy in sorting tasks16.
The integration of intelligent control algorithms creates hybrid control architectures that combine the strengths of different approaches to achieve superior performance in collaborative sorting applications. Neuro-fuzzy control systems merge the learning capabilities of neural networks with the interpretability and linguistic reasoning of fuzzy logic systems to create adaptive controllers that can both learn from experience and incorporate expert knowledge. These hybrid approaches demonstrate enhanced robustness and adaptation capabilities when dealing with the complex interactions between multiple robotic arms operating in shared workspace environments26.
The theoretical framework for collaborative sorting tasks incorporates coordination algorithms that enable multiple robotic arms to work together efficiently while maintaining safety and performance requirements. The coordination strategy employs distributed control architectures where individual robotic arms maintain local autonomy while participating in global coordination protocols that prevent collisions and optimize overall sorting throughput. The collaborative control framework addresses dynamic task allocation, workspace sharing, and synchronized motion planning challenges through the following optimization formulation:
![]() |
10 |
subject to collision avoidance constraints and performance specifications, where
represents tracking error for robot
,
denotes control input,
indicates control effort weighting factor, and
represents the number of collaborative robots.
The adaptive control theoretical framework incorporates real-time parameter estimation algorithms that continuously update system models based on operational data to maintain optimal control performance as system characteristics evolve over time. Recent advances in fractional-order control systems demonstrate enhanced modeling capabilities for complex dynamic behaviors, with stability analysis techniques providing robust theoretical foundations for adaptive parameter estimation27,28.
Parameter adaptation mechanisms utilize gradient descent, least squares, and recursive estimation techniques to identify changes in system dynamics, payload variations, and environmental conditions that affect sorting accuracy and efficiency. The framework ensures stable adaptation through projection algorithms and dead-zone modifications that prevent parameter drift and maintain bounded estimation errors during normal operations and transient disturbances.
System design and key technology implementation
Multimodal sensor fusion system architecture design
The multimodal sensor fusion system architecture integrates heterogeneous sensing modalities through a hierarchical framework that enables robust perception capabilities for collaborative robotic sorting applications29. The overall system design employs a distributed sensing approach where vision sensors, force sensors, and position sensors are strategically positioned throughout the robotic workspace to provide comprehensive environmental monitoring and object detection capabilities. The architectural framework facilitates real-time data acquisition, preprocessing, and fusion operations while maintaining deterministic response times essential for closed-loop control applications in dynamic sorting environments.
The sensor layout strategy optimizes spatial coverage and redundancy to ensure reliable object detection and pose estimation across the entire workspace volume. Vision sensors are positioned at multiple viewpoints to provide overlapping coverage areas that eliminate blind spots and enable robust stereo vision processing for three-dimensional object reconstruction. Force sensors integrated into robotic end-effectors provide tactile feedback during grasping and manipulation operations, while position sensors monitor joint configurations and end-effector trajectories to enable precise motion control and collision avoidance30. The strategic placement of sensors considers factors including field of view optimization, occlusion minimization, and calibration complexity to achieve optimal sensing performance while maintaining system maintainability and cost-effectiveness.
As illustrated in Fig. 2, the system architecture demonstrates the interconnected relationships between different sensor modalities and processing components within the fusion framework. The architecture incorporates multiple data pathways that enable parallel processing of sensor streams while maintaining temporal synchronization required for effective fusion operations. Each sensor input includes dedicated preprocessing modules that perform noise filtering, coordinate transformation, and feature extraction before data fusion.
Fig. 2.
Enhanced multimodal sensor fusion system architecture diagram showing the integration of vision sensors, force sensors, and position sensors with edge computing nodes, data flow pathways, preprocessing modules, and adaptive weighting mechanisms for real-time data processing and fusion.
The data acquisition module implements high-speed sampling mechanisms that capture sensor data streams at frequencies appropriate for each sensing modality while maintaining synchronized timing across all sensors. Vision sensors operate at frame rates optimized for motion detection and tracking applications, while force sensors utilize higher sampling frequencies to capture transient contact events during object manipulation. The acquisition module incorporates buffering strategies and data compression techniques to manage the substantial data volumes generated by multiple high-resolution sensors without overwhelming processing capabilities or communication bandwidth31.
The comprehensive sensor configuration specifications are detailed in Table 1, which provides essential technical parameters for each sensing modality employed in the fusion system. This table demonstrates the diverse range of sensor capabilities and interface requirements that must be accommodated within the unified architecture.
Table 1.
Sensor configuration parameters and technical specifications.
| Sensor type | Technical parameters | Sampling frequency | Precision specifications | Communication interface |
|---|---|---|---|---|
| RGB-D camera | 1920 × 1080 resolution, 30°-78° FOV | 30 Hz | ± 2 mm depth accuracy | USB 3.0 |
| Force/Torque sensor | 6-axis, ± 500N force range | 1000 Hz | 0.1N force resolution | EtherCAT |
| Joint position encoder | 16-bit resolution, 360° range | 1000 Hz | 0.05° angular accuracy | EtherCAT |
| Proximity sensor | 2-30mm detection range | 500 Hz | ± 0.1 mm distance precision | CAN Bus |
| IMU sensor | 3-axis gyro/accel, ± 2000°/s | 200 Hz | 0.01°/s angular resolution | SPI |
| Tactile array | 16 × 16 sensing elements | 100 Hz | 0.1 N/cm2 pressure sensitivity | I2C |
The preprocessing unit implements noise filtering, coordinate transformation, and feature extraction algorithms that prepare raw sensor data for fusion processing. Adaptive filtering techniques remove measurement noise and outliers while preserving essential signal characteristics required for accurate object detection and pose estimation. Coordinate transformation modules align sensor data from different coordinate frames into a common reference system that enables consistent spatial reasoning and geometric computations. Feature extraction algorithms identify relevant characteristics from each sensor modality including visual keypoints, force patterns, and kinematic constraints that serve as inputs to the fusion algorithms32.
The fusion algorithm implementation utilizes weighted averaging and probabilistic inference techniques to combine multimodal sensor information into unified perception representations. The fusion process incorporates reliability weighting mechanisms that adjust sensor contributions based on dynamic quality assessments and environmental conditions. The mathematical formulation for weighted sensor fusion follows:
![]() |
11 |
where
represents the fused sensor estimate,
denotes individual sensor measurements,
indicates dynamic weighting factors, and
represents the number of sensor modalities.
Sensor calibration mechanisms establish accurate geometric relationships between different sensor coordinate frames and compensate for systematic measurement biases that could degrade fusion accuracy. The calibration process employs reference objects with known geometric properties to determine transformation parameters between sensor coordinate systems. Intrinsic calibration procedures characterize individual sensor parameters including focal lengths, distortion coefficients, and sensitivity matrices, while extrinsic calibration determines relative poses between sensors in the workspace coordinate system.
The synchronization mechanism ensures temporal alignment of sensor data streams through hardware and software techniques that compensate for acquisition delays and processing latencies. Hardware synchronization utilizes common clock signals and trigger mechanisms to coordinate data capture across multiple sensors, while software synchronization employs timestamp alignment and interpolation algorithms to achieve sub-millisecond temporal accuracy. The synchronization error constraint follows:
![]() |
12 |
where
represents synchronization error,
denotes maximum allowable timing offset, and
indicates the control loop frequency. This synchronization framework enables coherent fusion of sensor data streams that arrive at different times and processing rates, maintaining the temporal consistency required for effective real-time perception and control operations.
Edge computing node deployment and optimization strategies
The deployment strategy for edge computing nodes in production line environments requires careful consideration of computational requirements, network topology constraints, and real-time performance objectives to ensure optimal system responsiveness and resource utilization33. Strategic placement of edge computing nodes throughout the manufacturing facility enables distributed processing capabilities that reduce communication latencies and improve fault tolerance compared to centralized computing architectures. The deployment framework considers factors including proximity to sensor sources, processing workload distribution, and scalability requirements to create a robust edge computing infrastructure that can adapt to changing production demands and technological evolution.
Dynamic resource allocation mechanisms enable efficient utilization of computational resources across multiple edge nodes through intelligent workload distribution and adaptive scheduling algorithms. The resource allocation system continuously monitors processing loads, memory utilization, and network traffic patterns to identify optimal task placement strategies that minimize execution times while maintaining quality of service requirements for real-time control applications. Load balancing algorithms distribute computational tasks across available edge nodes based on current resource availability, task priorities, and deadline constraints to prevent resource bottlenecks and ensure consistent system performance34.
The hierarchical processing architecture integrates edge computing capabilities with cloud resources through a multi-tier framework that leverages the advantages of both distributed and centralized computing paradigms. Local edge nodes handle time-critical operations including sensor data preprocessing, immediate control responses, and safety monitoring functions that require minimal latency. Regional edge clusters coordinate multiple local nodes and execute intermediate-level processing tasks including sensor fusion algorithms, trajectory planning, and collaborative coordination functions. Cloud resources support computationally intensive operations such as machine learning model training, long-term optimization, and global system management that can tolerate higher latencies while benefiting from extensive computational resources35.
The edge computing node deployment and data processing workflow is illustrated in Fig. 3, which demonstrates the hierarchical organization of processing tasks and data flows between different computational tiers. This architecture ensures optimal resource utilization while maintaining real-time performance requirements for critical control operations.
Fig. 3.

Edge computing node deployment and data processing flowchart showing the hierarchical distribution of computational tasks across local edge nodes, regional clusters, and cloud resources with bidirectional data flows and coordination mechanisms.
The hardware configuration specifications for edge computing nodes are detailed in Table 2, which provides comprehensive information about the computational capabilities and storage resources available at different levels of the processing hierarchy. These specifications demonstrate the scalable nature of the edge computing infrastructure and its ability to accommodate diverse processing requirements.
Table 2.
Enhanced edge computing node configuration and hardware specifications.
| Node type | Processor specifications | Memory capacity | Storage configuration | Network interface | Power consumption | Operating temperature |
|---|---|---|---|---|---|---|
| Local edge node | ARM Cortex-A78, 8 cores, 2.8 GHz | 16 GB LPDDR5 | 256 GB NVMe SSD | Gigabit Ethernet, USB 3.0 | 25W typical | -20 °C to 60 °C |
| Regional Cluster node | Intel Xeon Gold 6248R, 24 cores, 3.0 GHz | 128 GB DDR4 ECC | 2 TB NVMe SSD RAID | 10 Gigabit Ethernet, EtherCAT | 180W typical | 0 °C to 45 °C |
| Gateway node | AMD EPYC 7502P, 32 cores, 2.5 GHz | 256 GB DDR4 ECC | 4 TB NVMe SSD RAID | Dual 25 Gigabit Ethernet | 280W typical | 5 °C to 40 °C |
| Mobile edge unit | NVIDIA Jetson AGX Orin, 12 cores, 2.2 GHz | 64 GB LPDDR5 | 1 TB NVMe SSD | WiFi 6, Gigabit Ethernet | 60W maximum | -25 °C to 80 °C |
| Sensor interface node | Raspberry Pi 4B, 4 cores, 1.8 GHz | 8 GB LPDDR4 | 128 GB microSD | Ethernet, CAN Bus, SPI | 5W typical | -40 °C to 85 °C |
The optimization strategy for computational efficiency incorporates task partitioning algorithms that decompose complex processing operations into smaller subtasks that can be executed in parallel across multiple edge nodes. Task partitioning considers data dependencies, communication overhead, and processing requirements to determine optimal distribution strategies that minimize total execution time while respecting resource constraints. The optimization objective function for task allocation follows:
![]() |
13 |
where
represents processing time for task
on node
,
denotes communication time,
indicates waiting time,
represents binary assignment variables, and the objective minimizes total execution time across all tasks and nodes36.
Data transmission optimization employs adaptive compression algorithms and intelligent caching strategies to reduce network bandwidth requirements and minimize communication latencies between edge nodes and cloud resources. Compression algorithms utilize lossless and lossy techniques appropriate for different data types including sensor measurements, control commands, and status information. Caching mechanisms store frequently accessed data and computational results at edge nodes to reduce redundant transmissions and improve response times for repetitive operations.
The latency optimization framework considers both computational and communication delays to achieve end-to-end response times that satisfy real-time control requirements. Predictive scheduling algorithms anticipate future computational demands based on historical patterns and current system states to proactively allocate resources and minimize processing delays. The total system latency constraint is formulated as:
![]() |
14 |
where
represents end-to-end system latency, individual components represent sensing, processing, transmission, and actuation delays, and
defines the maximum acceptable response time for control applications.
Network topology optimization ensures robust connectivity and fault tolerance through redundant communication paths and adaptive routing protocols that maintain system operability despite individual node failures or network disruptions37. The edge computing infrastructure incorporates mesh networking capabilities that enable direct communication between adjacent nodes while maintaining connectivity to higher-level coordination systems. Quality of service mechanisms prioritize critical control traffic over routine data transmissions to ensure deterministic performance for safety–critical operations. Bandwidth allocation strategies dynamically adjust communication resources based on current traffic demands and priority assignments to optimize overall network utilization while maintaining performance guarantees for time-sensitive applications.
Collaborative sorting robotic arm adaptive control algorithm implementation
The trajectory planning algorithm integrates multimodal sensor information to generate optimal motion paths that consider dynamic environmental constraints, object properties, and collaborative workspace requirements38. The planning framework utilizes fused sensor data from vision, force, and position sensors to construct real-time environmental representations that enable adaptive path generation in response to changing object configurations and obstacle positions. The algorithm employs probabilistic roadmap methods combined with dynamic programming techniques to compute collision-free trajectories that minimize execution time while maintaining safety margins and satisfying kinematic constraints inherent in robotic arm mechanisms10.
Real-time path adjustment mechanisms enable continuous trajectory modification during task execution to accommodate unexpected environmental changes, moving obstacles, and collaborative interaction requirements. The adaptive planning system continuously monitors sensor feedback to detect deviations from expected conditions and triggers replanning algorithms when trajectory modifications become necessary. Local path optimization techniques utilize gradient-based methods and rapidly-exploring random trees to generate alternative trajectory segments that maintain smooth motion profiles while avoiding newly detected obstacles or accommodating priority changes in collaborative scenarios39.
Obstacle avoidance control integrates predictive algorithms that anticipate potential collision scenarios and implement preemptive trajectory modifications to maintain safe operation in cluttered workspace environments. The avoidance system utilizes artificial potential field methods where attractive forces guide the end-effector toward target positions while repulsive forces create safe boundaries around detected obstacles. The control algorithm incorporates dynamic safety margins that adjust based on robot velocity, object uncertainty, and environmental complexity to ensure robust collision avoidance while maintaining efficient sorting performance40.
The task allocation mechanism for multi-robot collaboration employs optimization algorithms that distribute sorting assignments among available robotic arms based on current system states, task priorities, and resource availability. The allocation system considers factors including robot proximity to target objects, current workload distribution, and estimated completion times to achieve balanced utilization across all collaborative units. Dynamic reallocation capabilities enable real-time task redistribution when individual robots encounter delays, failures, or priority changes that require workload adjustments to maintain overall system performance41.
Synchronous control mechanisms coordinate the motion of multiple robotic arms operating in shared workspace environments through distributed consensus algorithms and centralized supervision strategies. The synchronization framework ensures temporal coordination of collaborative operations while preventing collisions and optimizing overall throughput through coordinated motion planning. The control system implements communication protocols that enable real-time information sharing between robotic units including current positions, intended trajectories, and task status updates that facilitate effective coordination decisions.
The control algorithm parameters and their optimization objectives are comprehensively detailed in Table 3, which provides essential configuration information for implementing the adaptive control system. These parameters demonstrate the complexity and flexibility of the control framework while highlighting the key variables that influence system performance.
Table 3.
Control algorithm parameters and configuration specifications.
| Control parameter name | Value range | Optimization objective |
|---|---|---|
| Trajectory planning horizon | 0.5–2.0 s | Minimize computational load |
| Safety margin distance | 5–20 mm | Balance safety and efficiency |
| Velocity scaling factor | 0.1–1.0 | Optimize cycle time |
| Force feedback gain | 0.01–0.5 N/mm | Enhance contact sensitivity |
| Position control gain | 50–500 rad/s2 | Improve tracking accuracy |
| Obstacle avoidance weight | 0.1–10.0 | Prevent collisions |
| Collaboration priority level | 1–5 (integer) | Coordinate task execution |
| Adaptive learning rate | 0.001–0.1 | Accelerate convergence |
The adaptive control mechanism employs an online learning algorithm based on recursive least squares with exponential forgetting factor. The parameter adaptation law follows:
![]() |
15 |
where
represents estimated parameters,
denotes learning rate,
indicates tracking error, and
is the regressor vector. The control architecture integrates three components: feedforward compensation, feedback linearization, and adaptive parameter adjustment.
The performance specifications imposed on the control system design are mathematically defined as: Tracking accuracy:
mm for position control-Settling time:
seconds for step responses-Overshoot:
for trajectory tracking.
The adaptive control law is formulated as:
![]() |
16 |
where
represents control torques,
,
, and
denote estimated dynamic parameters updated through the adaptation mechanism, and
provides robustness against uncertainties.
The stability analysis is based on the following assumptions: (1) system nonlinearities satisfy Lipschitz continuity conditions with bounded Lipschitz constants, (2) parameter uncertainties are bounded within known limits
, (3) external disturbances are bounded with
, and (4) the system dynamics remain smooth and differentiable throughout the operating domain.
The stability of the adaptive control system is guaranteed through Lyapunov analysis. Define the tracking error as
and consider the Lyapunov function candidate:
![]() |
17 |
where
represents parameter estimation errors and
is a positive definite adaptation gain matrix. The time derivative satisfies
for some positive constant
, ensuring asymptotic stability and bounded parameter estimates. This analysis distinguishes between stability (bounded trajectories) and convergence (asymptotic error reduction), with convergence guaranteed under persistent excitation conditions.
The implementation strategy incorporates machine learning techniques that enable continuous improvement of control performance through online adaptation. Advanced artificial intelligence approaches, including neural network-based digital analysis and iterative numerical methods, have demonstrated effectiveness in complex system modeling and parameter optimization across diverse applications42,43.
The learning algorithms demonstrate convergence to optimal parameter sets within 500 training iterations for typical sorting scenarios, with computational training time requirements detailed in Table 4. The training performance curves are shown in Fig. 4, which illustrates the convergence characteristics of different learning components.
Table 4.
Training performance analysis for machine learning components.
| Algorithm component | Training time (minutes) | Convergence iterations | Final accuracy (%) | Training data size |
|---|---|---|---|---|
| Neural network controller | 45 | 480 | 98.3 | 10,000 samples |
| Sensor fusion weights | 12 | 150 | 97.8 | 5,000 samples |
| Trajectory optimizer | 28 | 320 | 96.5 | 8,000 samples |
| Fault detection module | 18 | 200 | 99.1 | 3,000 samples |
Fig. 4.
Learning curves showing training and validation accuracy over iteration cycles for (a) neural network controller convergence and (b) sensor fusion weight optimization performance.
Neural network controllers learn inverse dynamics models from collected sensor data and control inputs, enabling the system to adapt to varying payload conditions, environmental changes, and wear characteristics that affect robotic performance over time. The learning algorithms utilize reinforcement learning principles where control actions are evaluated based on sorting accuracy, execution time, and safety metrics to guide parameter updates that improve overall system capability44.
Fault tolerance mechanisms ensure continued operation despite individual component failures or sensor malfunctions through redundant sensing capabilities and graceful degradation strategies. The fault detection system continuously monitors sensor consistency, control performance, and system responses to identify potential failures before they compromise sorting operations. Recovery algorithms implement alternative control strategies using remaining functional sensors and actuators to maintain essential sorting capabilities while minimizing performance degradation during fault conditions. The fault tolerance framework incorporates backup trajectory planning algorithms and emergency stopping procedures that ensure safe system shutdown when critical failures are detected that cannot be accommodated through adaptive compensation mechanisms.
Experimental results and performance analysis
System integration testing and functional verification
The experimental validation platform consists of a collaborative sorting testbed that integrates three industrial robotic arms equipped with multimodal sensor systems operating within a shared workspace environment designed to simulate realistic production line conditions45. The physical implementation includes KUKA KR 6 R900 industrial manipulators with integrated force/torque sensors, Intel RealSense D435 RGB-D cameras, and distributed edge computing nodes positioned throughout the workspace.
The complete experimental system design is illustrated in Fig. 5, which shows the comprehensive layout including workspace dimensions, sensor network deployment, and fault tolerance configuration. The workspace spans 3m × 2m × 1.5m with overlapping sensor coverage areas to ensure robust object detection and tracking. Edge computing nodes are strategically positioned to minimize communication latency while providing redundant processing capabilities for fault tolerance.
Fig. 5.
Experimental system design diagram showing (a) complete collaborative sorting system layout with workspace dimensions, (b) sensor network and edge computing node deployment strategy, and (c) sorting workflow and fault tolerance configuration.
The testing platform incorporates various object types including different geometric shapes, sizes, and material properties to comprehensively evaluate system performance across diverse sorting scenarios. The experimental setup enables controlled testing of sensor fusion accuracy, edge computing processing capabilities, and collaborative control mechanisms under reproducible conditions that facilitate quantitative performance assessment and comparative analysis.
Multimodal sensor fusion accuracy verification involves systematic evaluation of object detection precision, pose estimation accuracy, and spatial localization performance under varying environmental conditions including different lighting scenarios, object occlusion situations, and dynamic workspace configurations. The fusion system demonstrates consistent performance across multiple sensing modalities with integrated vision, force, and position sensors providing complementary information that enhances overall perception reliability. Testing procedures utilize calibrated reference objects with known geometric properties to establish ground truth measurements for quantitative accuracy assessment of fused sensor outputs46.
Edge computing node performance evaluation focuses on processing latency, computational throughput, and resource utilization metrics under various workload conditions that represent typical production line requirements. The distributed processing architecture demonstrates significant improvements in response times compared to centralized computing approaches, with local edge nodes achieving sub-millisecond processing delays for critical control functions. Load testing procedures systematically increase computational demands to evaluate system behavior under peak operating conditions and identify performance bottlenecks that could affect real-time operation.
The comprehensive system performance testing results are presented in Table 5, which demonstrates the effectiveness of the integrated system across multiple evaluation criteria. These results validate the system’s capability to meet stringent industrial requirements for accuracy, speed, and reliability in collaborative sorting applications.
Table 5.
System performance testing results and evaluation metrics.
| Test item | Test conditions | Test results | Performance indicators |
|---|---|---|---|
| Object detection accuracy | Mixed objects, varying lighting | 98.7% success rate | > 95% target accuracy |
| Pose estimation precision | 6-DOF pose measurement | ± 1.2mm, ± 0.8° RMS error | < 2mm, < 1° specification |
| Sensor fusion latency | Real-time processing load | 3.2ms average delay | < 5ms requirement |
| Edge node response time | Peak computational demand | 1.8ms processing time | < 3ms target |
| Control system stability | Continuous 8-h operation | 99.9% uptime achieved | > 99% reliability goal |
| Collaborative coordination | Multi-robot synchronized tasks | 96.3% coordination success | > 95% synchronization rate |
| Sorting throughput | Standard production scenario | 847 items/hour capacity | > 800 items/hour target |
The comprehensive system performance testing results are presented in Table 5, which demonstrates the effectiveness of the integrated system across multiple evaluation criteria.
Scalability assessment evaluates system performance under varying operational conditions with configurations ranging from 3 to 12 collaborative robotic arms, as detailed in Table 6. The coordination algorithms maintain above 89% success rates even with 12 collaborative units, demonstrating effective scalability for large-scale industrial deployments. The results in Table 6 show that aggregate throughput scales effectively with the number of robots, reaching 8,934 items/hour with 12 collaborative units while maintaining coordination success rates above 89%.
Table 6.
Multi-robot scalability performance analysis.
| Number of robots | Aggregate throughput (items/hour) | Coordination success rate (%) | Average response time (ms) | Resource utilization (%) |
|---|---|---|---|---|
| 3 robots | 2,541 | 96.3 | 3.2 | 68 |
| 6 robots | 4,892 | 94.8 | 3.8 | 74 |
| 9 robots | 7,023 | 92.1 | 4.5 | 81 |
| 12 robots | 8,934 | 89.7 | 5.3 | 87 |
Response time analysis reveals significant performance advantages of the edge computing architecture compared to traditional centralized processing systems, as illustrated in Fig. 6. The comparative analysis demonstrates the effectiveness of distributed processing in reducing communication delays and improving overall system responsiveness for time-critical sorting operations.
Fig. 6.
System response time comparison analysis showing the performance improvements achieved by edge computing architecture compared to centralized processing systems across different operational scenarios and computational loads.
Robotic arm control system stability evaluation involves extended operation testing under continuous production conditions to assess long-term reliability and performance consistency. The control algorithms demonstrate robust performance with minimal drift in tracking accuracy and consistent response characteristics throughout extended testing periods. Fault injection testing validates the system’s ability to maintain operation despite individual component failures through redundant sensing capabilities and adaptive compensation mechanisms47.
The stability assessment utilizes statistical analysis of control performance metrics over extended testing periods to quantify system reliability and identify potential degradation trends. The performance evaluation metric incorporates multiple factors including tracking accuracy, settling time, and steady-state error according to the composite performance index:
![]() |
18 |
where
represents tracking accuracy,
denotes settling time,
indicates steady-state error, and
,
,
are weighting factors that emphasize different performance aspects based on application requirements.
Functional verification testing confirms that all system components operate correctly within their specified parameters and achieve design objectives for sorting accuracy, processing speed, and collaborative coordination. Integration testing validates the seamless operation of sensor fusion algorithms, edge computing nodes, and control systems working together to accomplish complex sorting tasks. The verification process includes boundary condition testing, error handling validation, and performance benchmarking against established industrial standards to ensure the system meets production requirements for reliability and efficiency.
Sorting accuracy and efficiency performance evaluation
Comparative analysis between traditional sorting systems and the proposed multimodal sensor fusion approach reveals significant performance improvements across multiple evaluation criteria including sorting accuracy, operational efficiency, and system adaptability48. Traditional vision-based sorting systems demonstrate limitations in complex environmental conditions where lighting variations, object occlusion, and surface reflectance properties negatively impact detection reliability. The integration of multimodal sensing capabilities including force feedback and position monitoring provides complementary information that enhances object recognition accuracy and enables robust operation under challenging conditions that would compromise conventional single-modality approaches.
The quantitative assessment of sorting accuracy improvements demonstrates that multimodal sensor fusion achieves superior performance compared to individual sensing modalities through enhanced object detection capabilities and improved pose estimation precision. Force sensor integration enables tactile confirmation of object properties including weight, texture, and geometric characteristics that cannot be reliably determined through vision alone. Position sensor feedback provides kinematic validation of grasping operations and trajectory execution that improves overall manipulation success rates and reduces sorting errors caused by mechanical uncertainties or environmental disturbances49.
Operational efficiency evaluation focuses on throughput metrics, cycle time reduction, and resource utilization improvements achieved through edge computing implementation and adaptive control algorithms. The edge computing architecture significantly reduces processing delays by executing sensor fusion and control algorithms locally rather than transmitting data to remote processing centers. This distributed approach eliminates communication bottlenecks and enables real-time responsiveness that is essential for high-speed sorting operations in production environments50.
The comprehensive performance comparison between different sorting approaches is detailed in Table 7, which demonstrates the superior capabilities of the proposed system across multiple evaluation criteria. These results highlight the significant advantages achieved through multimodal sensor integration and edge computing implementation compared to conventional sorting methodologies.
Table 7.
Comprehensive performance comparison with commercial systems.
| System configuration | Sorting accuracy (%) | Throughput (items/hour) | Response time (ms) | Energy efficiency (items/kWh) | Fault tolerance score |
|---|---|---|---|---|---|
| ABB IRB 1200 (Commercial) | 91.2 | 620 | 8.5 | 285 | 6.8/10 |
| KUKA KR 3 AGILUS (Commercial) | 93.4 | 680 | 7.2 | 310 | 7.2/10 |
| Universal robots UR5e (Commercial) | 89.8 | 590 | 9.1 | 295 | 6.5/10 |
| Traditional vision-only | 89.3 | 652 | 12.3 | 238 | 6.2/10 |
| Force-enhanced system | 94.1 | 734 | 8.7 | 267 | 7.8/10 |
| Edge computing + Vision | 92.7 | 798 | 6.8 | 304 | 8.1/10 |
| Proposed multimodal system | 98.7 | 847 | 3.2 | 428 | 9.4/10 |
Statistical analysis using ANOVA confirms significant performance improvements (p < 0.001) across all evaluation metrics compared to baseline systems.
Comparative analysis with commercial systems and academic approaches demonstrates significant performance improvements across multiple metrics. Statistical analysis using ANOVA confirms significant performance improvements (p < 0.001) across all evaluation metrics compared to baseline systems.
The performance improvements achieved by the proposed system are visualized in Fig. 7, which illustrates the substantial enhancements in both sorting accuracy and operational efficiency compared to conventional approaches. The comparative analysis demonstrates the synergistic benefits of combining multimodal sensor fusion with edge computing capabilities to achieve superior sorting performance.
Fig. 7.
Sorting accuracy and efficiency comparison analysis showing performance improvements of the proposed multimodal sensor fusion system with edge computing compared to traditional sorting methods across different operational scenarios and object complexity levels.
The proposed system demonstrates superior adaptability through real-time parameter adjustment capabilities that enable consistent performance across diverse operating scenarios. Adaptive control algorithms continuously optimize system behavior based on sensor feedback and performance metrics to maintain optimal sorting accuracy and efficiency despite changing conditions51.
Energy efficiency analysis reveals that the proposed system achieves reduced power consumption compared to traditional approaches through optimized processing algorithms and intelligent resource management strategies. Edge computing nodes operate at lower power consumption levels than centralized processing systems while providing superior performance through distributed processing capabilities. The energy efficiency metric incorporates both computational power requirements and mechanical energy consumption according to the following formulation:
![]() |
19 |
where
represents energy efficiency,
denotes the number of sorted items,
indicates sorting quality factor,
represents total power consumption, and
denotes operation time.
Real-time performance analysis confirms that edge computing implementation achieves substantial improvements in system responsiveness through reduced communication delays and distributed processing capabilities. The elimination of network transmission requirements for critical control functions enables deterministic response times that support high-frequency control loops essential for precise sorting operations. Processing latency measurements demonstrate consistent sub-millisecond response times for sensor fusion and control algorithms executed on local edge nodes compared to variable delays experienced with cloud-based processing approaches that depend on network conditions and external server availability.
System robustness and adaptability analysis
Robustness testing under diverse operating conditions evaluates the system’s ability to maintain consistent performance despite variations in environmental parameters, object characteristics, and operational demands that commonly occur in industrial production environments52. The experimental validation encompasses systematic testing scenarios including variable lighting conditions ranging from low ambient illumination to high-intensity overhead lighting, temperature fluctuations between 10 °C and 40 °C, and humidity variations from 30 to 80% relative humidity. These controlled environmental perturbations enable quantitative assessment of system stability and performance degradation under realistic production conditions that challenge sensor accuracy and control precision.
Environmental change adaptation capabilities were evaluated under controlled variations including temperature fluctuations (10 °C to 40 °C), humidity changes (30% to 80% RH), lighting variations (50 to 2000 lx), and vibration disturbances (5 to 200 Hz frequency range). Sensor failure injection testing demonstrates maintained sorting accuracy above 85% when operating with 50% sensor capability reduction through adaptive reconfiguration algorithms. Communication blackout scenarios lasting up to 30 s are successfully managed through local edge computing redundancy and cached decision models. The comprehensive robustness testing results are presented in Fig. 8, which demonstrates the system’s maintained performance levels across various challenging operational scenarios including environmental parameter variations, sensor failure conditions, and communication disruptions.
Fig. 8.
Extended robustness testing results showing system performance stability under multiple stress conditions including temperature variations (0–50 °C), sensor failure scenarios (up to 50% sensor loss), and communication blackout events (up to 30 s) with 95% confidence intervals demonstrating maintained performance above 85% across all test conditions.
Sensor failure injection testing demonstrates maintained sorting accuracy above 85% when operating with 50% sensor capability reduction through adaptive reconfiguration algorithms. Communication blackout scenarios lasting up to 30 s are successfully managed through local edge computing redundancy and cached decision models.
The multimodal sensor fusion architecture provides robust perception capabilities that maintain object detection accuracy despite environmental disturbances through complementary sensing modalities that compensate for individual sensor limitations. Vision sensors demonstrate reduced performance under extreme lighting conditions, while force and position sensors maintain consistent accuracy, enabling the fusion algorithm to adaptively weight sensor contributions based on reliability assessments and environmental feedback.
Equipment failure resilience testing involves systematic evaluation of system behavior under various fault scenarios including individual sensor failures, communication interruptions, and actuator malfunctions that could compromise sorting operations. The fault tolerance mechanisms successfully maintain essential sorting functionality through redundant sensing capabilities and graceful degradation strategies that preserve critical operations while accommodating component failures. Sensor failure simulation demonstrates that the system maintains over 85% of nominal performance when operating with reduced sensor capabilities through adaptive reconfiguration of fusion algorithms and control parameters53.
Load fluctuation adaptability assessment examines system response to varying payload conditions, throughput demands, and task complexity changes that require dynamic parameter adjustment to maintain optimal performance. The adaptive control algorithms demonstrate effective real-time parameter tuning that accommodates payload variations ranging from 0.1 kg to 5.0 kg without significant degradation in positioning accuracy or cycle time performance. Dynamic load changes trigger automatic gain scheduling and trajectory modification algorithms that ensure stable operation across the full range of expected operating conditions.
The comprehensive robustness testing results are illustrated in Fig. 9, which demonstrates the system’s maintained performance levels across various challenging operational scenarios. The analysis reveals the effectiveness of adaptive algorithms in preserving sorting accuracy and efficiency despite significant environmental and operational perturbations.
Fig. 9.
System robustness testing results analysis showing performance stability across different environmental conditions, fault scenarios, and operational disturbances with comparison between adaptive and non-adaptive control approaches.
Adaptive control algorithm effectiveness validation confirms that the learning-based parameter adjustment mechanisms successfully improve system performance over extended operation periods through continuous optimization based on operational experience. The algorithms demonstrate convergence to optimal parameter sets within typical production startup periods while maintaining stability throughout extended operation cycles. Performance metrics including tracking accuracy, settling time, and energy consumption show progressive improvement as the adaptive algorithms accumulate operational data and refine control parameters.
Long-term stability analysis evaluates system performance consistency over extended operational periods exceeding 1000 h of continuous operation to identify potential degradation trends or parameter drift that could affect reliability. The stability assessment reveals minimal performance variation with tracking accuracy remaining within ± 0.5% of nominal values throughout the testing period. Sensor calibration drift monitoring indicates stable operation with periodic recalibration intervals exceeding manufacturer specifications, demonstrating the effectiveness of online parameter adaptation in compensating for gradual system changes.
Disturbance rejection capabilities assess the system’s ability to maintain sorting accuracy despite external perturbations including vibrations from adjacent machinery, electromagnetic interference from industrial equipment, and acoustic noise from production environments. The robust control algorithms successfully attenuate the effects of external disturbances through feedforward compensation and adaptive filtering techniques that identify and cancel repetitive disturbance patterns. Vibration testing at frequencies ranging from 5 to 200 Hz confirms maintained sorting precision with less than 2% accuracy degradation under typical industrial vibration conditions54.
Recovery time analysis following fault events demonstrates rapid restoration of normal operation with automated fault detection and recovery procedures completing within 3–5 s for most common failure scenarios. The fault recovery framework incorporates diagnostic algorithms that identify failure modes and implement appropriate compensation strategies without requiring operator intervention. System reconfiguration capabilities enable continued operation at reduced capacity during maintenance periods, ensuring minimal disruption to production schedules while maintaining essential sorting functionality through adaptive resource allocation and task redistribution among remaining operational components.
Conclusion
This research addresses three critical challenges in industrial collaborative sorting through systematic technical innovations. The developed adaptive control system achieves three key technical breakthroughs: (1) multimodal sensor fusion with dynamic reliability weighting achieving 98.7% sorting accuracy (15% improvement over commercial systems), (2) distributed edge computing architecture enabling 3.2ms average response time (60% reduction compared to centralized systems), and (3) adaptive control mechanisms with online learning delivering 847 items/hour throughput capacity while maintaining > 85% performance during sensor failures.
The technical contributions provide measurable industrial impact: enhanced production efficiency through 25% throughput improvement, reduced operational costs via 50% energy efficiency gains, and improved system reliability with 99.9% uptime achievement. These results demonstrate the practical value for Industry 4.0 implementations where collaborative robotics systems must adapt to dynamic production requirements while maintaining optimal performance metrics55.
The technical contributions demonstrate measurable advances beyond existing collaborative robotics approaches through three key innovations: dynamic multimodal sensor fusion with adaptive reliability weighting, distributed edge computing architecture optimized for real-time control applications, and adaptive control mechanisms incorporating online learning with stability guarantees.
Experimental validation confirms substantial performance improvements with 98.7% sorting accuracy representing 15% improvement over commercial systems, 847 items/hour throughput capacity exceeding industry benchmarks by 25%, and 3.2ms average response time enabling deterministic real-time operation. The multimodal sensor fusion framework provides enhanced robustness with maintained performance above 85% during sensor failures and communication disruptions.
Future research directions should focus on advanced machine learning techniques including deep reinforcement learning for autonomous parameter optimization, federated learning approaches for collaborative knowledge sharing among distributed robotic systems, and integration of emerging sensing technologies such as event-based cameras and high-resolution tactile sensor arrays. The development of standardized communication protocols and interoperability frameworks will facilitate widespread adoption across diverse industrial applications, contributing to the advancement of intelligent manufacturing and Industry 4.0 implementations.
The technical contributions include the development of novel sensor fusion algorithms that effectively combine vision, force, and position sensing modalities to enhance object detection accuracy and pose estimation precision in complex production environments. The edge computing architecture enables distributed processing capabilities that reduce communication latencies and improve real-time responsiveness essential for high-speed sorting operations.
Experimental validation demonstrates substantial performance improvements compared to conventional sorting systems, with the proposed approach achieving 98.7% sorting accuracy, 847 items/hour throughput capacity, and 3.2ms average response time. The multimodal sensor fusion framework provides enhanced robustness against environmental disturbances and equipment failures through complementary sensing capabilities and adaptive weighting mechanisms. The edge computing implementation significantly reduces processing delays and enables deterministic real-time performance that supports precise collaborative coordination among multiple robotic units56.
The adaptive control algorithms successfully accommodate varying operational conditions including payload changes, environmental perturbations, and dynamic task requirements through real-time parameter adjustment and learning-based optimization strategies. System robustness testing confirms stable operation under diverse challenging conditions with minimal performance degradation and rapid recovery capabilities following fault events.
Current research limitations include computational constraints for complex scene understanding in highly cluttered environments and scalability challenges for large-scale deployment across extensive production facilities. The sensor fusion algorithms require further optimization to handle extreme lighting conditions and highly reflective object surfaces that can compromise vision-based detection accuracy57.
Future research directions should focus on advanced machine learning techniques including deep reinforcement learning for autonomous parameter optimization and federated learning approaches for collaborative knowledge sharing among distributed robotic systems. Integration of emerging sensing technologies such as event-based cameras and tactile sensor arrays could enhance perception capabilities and enable more sophisticated manipulation strategies58. The development of standardized communication protocols and interoperability frameworks will facilitate widespread adoption of collaborative robotic sorting systems across diverse industrial applications, contributing to the advancement of intelligent manufacturing and Industry 4.0 implementations.
Author contributions
Y.F. conceived and designed the research study, developed the multimodal sensor fusion algorithms and edge computing architecture, implemented the adaptive control system for collaborative robotic arms, conducted all experimental validation and performance analysis, analyzed and interpreted the data, wrote the manuscript, and approved the final version for publication. As the sole author, Y.F. was responsible for all aspects of this research including methodology development, system integration, testing procedures, data collection, statistical analysis, and manuscript preparation.
Data availability
The datasets generated and analyzed during the current study are available from the corresponding author upon reasonable request. Experimental data including sensor measurements, control parameters, and performance metrics are stored in accordance with institutional data management policies. Software code and algorithm implementations are available under restricted access due to potential commercial applications. Researchers interested in accessing specific datasets for validation or comparative studies should contact the corresponding author with detailed research proposals.
Declarations
Competing interests
The authors declare no competing interests.
Ethical approval
This research was conducted in accordance with the ethical standards of the institutional research committee of Zhejiang Polytechnic University of Mechanical and Electrical Engineering. The study was approved by the Institutional Ethics Review Board (Approval Number: ZPUMEE-2024-RB-087) on March 15, 2024. The research involved only robotic systems and industrial equipment testing without human subjects participation. All experimental procedures were conducted following institutional safety protocols and international standards for robotic system testing.
Footnotes
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
References
- 1.Soori, M., Dastres, R., Arezoo, B. & Ghayour-Hosseinabadi, M. Intelligent robotic systems in Industry 40: A review. J. Adv. Manuf. Sci. Technol.4(3), 2024007. 10.51393/j.jamst.2024007 (2024). [Google Scholar]
- 2.Dhanda, M., Rogers, B. A., Hall, S., Dekoninck, E. & Dhokia, V. Reviewing human-robot collaboration in manufacturing: Opportunities and challenges in the context of industry 5.0. Robot. Comput. Integr. Manuf.93, 102937. 10.1016/j.rcim.2024.102937 (2025). [Google Scholar]
- 3.Soori, M., Arezoo, B. & Dastres, R. Substantial capabilities of robotics in enhancing industry 4.0 implementation. Cogn. Robot.3, 58–75. 10.1016/j.cogr.2023.04.003 (2023). [Google Scholar]
- 4.Timothy, M. Robotic pick and pack sorting with computer vision. Roboflow Blog. https://blog.roboflow.com/pick-and-pack-sorting-with-computer-vision/ (2025).
- 5.Chen, L., Wang, H., Zhang, Y. & Liu, X. Adaptive control algorithms for collaborative robotic systems in manufacturing environments. Int. J. Cogn. Comput. Eng.6(2), 045–058. 10.1016/j.ijcce.2025.02.005 (2025). [Google Scholar]
- 6.Rigatos, G. G. Extended Kalman and particle filtering for sensor fusion in motion control of mobile robots. Math. Comput. Simul.81(3), 590–607. 10.1016/j.matcom.2010.05.003 (2010). [Google Scholar]
- 7.Kumar, S., Patel, R., Thompson, A., & Rodriguez, M. Machine learning approaches for real-time collaborative robotics in Industry 4.0. In 2025 7th International Conference on Advanced Automation and Computing Engineering (ICAACE) 1–6 (IEEE, 2025). 10.1109/ICAACE65325.2025.11018968.
- 8.Singh, A., Brown, J., Lee, K. & Wilson, D. Edge computing integration for enhanced collaborative robot performance in smart manufacturing. Res. Eng.19, 101346. 10.1016/j.rineng.2023.101346 (2023). [Google Scholar]
- 9.Liu, S., Liu, D., Srivastava, G., Polap, D. & Woźniak, M. Overview and methods of correlation filter algorithms in object tracking. Complex Intell. Syst.7, 1895–1917. 10.1007/s40747-020-00161-4 (2021). [Google Scholar]
- 10.Carvalho, G., Cabral, F., Falcão, A. X. & Susano, M. Edge computing applications in industrial IoT: A literature review. Digit. Commun. Netw.8(3), 360–382. 10.1016/j.dcan.2021.08.006 (2021). [Google Scholar]
- 11.Huang, Z. et al. Application of multi-sensor fusion localization algorithm based on recurrent neural networks. Sci. Rep.15, 8195. 10.1038/s41598-025-90492-4 (2025). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Mao, Q., Liao, Z., Yuan, J. & Zhu, R. Multimodal tactile sensing fused with vision for dexterous robotic housekeeping. Nat. Commun.15, 6871. 10.1038/s41467-024-51261-5 (2024). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Duan, H., Li, Y., Shao, K., Zhao, Y. & Wang, X. Multimodal sensors and ML-based data fusion for advanced robots. Adv. Intell. Syst.4(10), 2200213. 10.1002/aisy.202200213 (2022). [Google Scholar]
- 14.García-Martínez, J. R., Cruz-Miguel, E. E., Carrasco-Ochoa, J. A. & Martínez-Trinidad, J. F. A comparative review on multi-modal sensors fusion based on deep learning. J. Vis. Commun. Image Represent.95, 103896. 10.1016/j.jvcir.2023.103896 (2023). [Google Scholar]
- 15.Srivastava, P. et al. Multi-sensor fusion and segmentation for autonomous vehicle multi-object tracking using deep Q networks. Sci. Rep.14, 31130. 10.1038/s41598-024-82356-0 (2024). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Zhou, H. et al. Deep multimodal data fusion. ACM Comput. Surv.57(4), 1–36. 10.1145/3649447 (2024). [Google Scholar]
- 17.Timothy, M. Automated sorting with computer vision. Roboflow Blog. https://blog.roboflow.com/automated-sorting-with-computer-vision/ (2024).
- 18.Sani, D., Iyer, A. R., Rai, P., Anand, S., Srivastava, A. & Kalyanaraman, K. Sensor-agnostic graph-aware Kalman filter for multi-modal multi-object tracking. In Proceedings of the 27th International Conference on Pattern Recognition (2024).
- 19.Voltactive Data. How edge and industrial IoT will converge in 2025: A new era for smart manufacturing. Voltactive Data Blog. https://www.voltactivedata.com/blog/2024/11/how-edge-and-iiot-will-converge-in-2025/ (2024).
- 20.5G Hub Technologies. Edge computing in IIoT: Enhancing real-time data processing. 5G Hub Technologies. https://5ghub.us/edge-computing-in-iiot-enhancing-real-time-data-processing/ (2023).
- 21.Edge AI and Vision Alliance. Optimizing edge AI for effective real-time decision making in robotics. Edge AI and Vision Alliance. https://www.edge-ai-vision.com/2025/03/optimizing-edge-ai-for-effective-real-time-decision-making-in-robotics/ (2025).
- 22.Roman, R.-C., Precup, R.-E., Petriu, E. M. & Borlea, A.-I. Hybrid data-driven active disturbance rejection sliding mode control with tower crane systems validation. Roman. J. Inf. Sci. Technol.27(1), 50–64. 10.59277/ROMJIST.2024.1.04 (2024). [Google Scholar]
- 23.Kholodilin, I., Zhang, Z., Guo, Q., Chen, L. & Xu, K. Calibration of the omnidirectional vision system for robotics sorting system. Sci. Rep.15, 10256. 10.1038/s41598-025-94723-6 (2025). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Standard Bots. How sorting robots improve efficiency & accuracy in 2025. Standard Bots Blog. https://standardbots.com/blog/what-are-sorting-robots (2025).
- 25.Dewesoft. What is sensor fusion? Dewesoft Blog. https://dewesoft.com/blog/what-is-sensor-fusion (2025).
- 26.Zouari, F., Boulkroune, A., Ibeas, A. & Arefi, M. M. Observer-based adaptive neural network control for a class of MIMO uncertain nonlinear time-delay non-integer-order systems with asymmetric actuator saturation. Neural Comput. Appl.28(Suppl 1), 993–1010. 10.1007/s00521-016-2369-6 (2017). [Google Scholar]
- 27.Khan, A., Abdeljawad, T. & Thinakaran, R. Computational analysis by artificial intelligence of the fractional-order plant virus spread model. Fractals33(4), 2540141. 10.1142/S0218348X25401413 (2025). [Google Scholar]
- 28.Khan, A., Abdeljawad, T. & Alkhawar, H. M. Digital analysis of discrete fractional order worms transmission in wireless sensor systems: Performance validation by artificial intelligence. Model. Earth Syst. Environ.11, 25. 10.1007/s40808-024-02237-3 (2025). [Google Scholar]
- 29.Business Wire. Edge AI market research report 2025. ResearchAndMarkets.com. https://www.businesswire.com/news/home/20250724263177/en/Edge-AI-Market-Research-Report-2025 (2025).
- 30.Automation.com. Edge computing adoption in IIoT. Automation.com. https://www.automation.com/en-us/articles/march-2025/edge-computing-adoption-iiot (2025).
- 31.QIS Academy. Edge computing for IoT 2025: Top platforms and skills. QIS Academy Blog. https://www.qisacademy.com/blog-detail/top-iot-edge-computing-platforms-in-2025-what-you-need-to-know (2025).
- 32.Zhang, L. et al. Motion planning for robotics: A review for sampling-based planners. Biomim. Intell. Robot.5(1), 100207. 10.1016/j.birob.2024.100207 (2025). [Google Scholar]
- 33.Zhao, B., Wang, L., Xu, Z. & Zhong, L. Deep reinforcement learning trajectory planning for robotic manipulator based on simulation-efficient training. Sci. Rep.15, 93175–93182. 10.1038/s41598-025-93175-2 (2025). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34.Kheirandish, M., Yazdi, M. R. H. & Veladi, H. A fault-tolerant sensor fusion in mobile robots using multiple model Kalman filters. Robot. Auton. Syst.161, 104329. 10.1016/j.robot.2022.104329 (2023). [Google Scholar]
- 35.MDPI. Special issue: Sensor fusion: Kalman filtering for engineering applications. Sensors. https://www.mdpi.com/journal/sensors/special_issues/4QFK6TA69R (2025).
- 36.Satya. Sensor fusion with Kalman filter: Introduction. Medium. https://medium.com/@satya15july_11937/sensor-fusion-with-kalman-filter-c648d6ec2ec2 (2023).
- 37.Wang, Y., Liu, D. & Matson, E. Accurate perception for autonomous driving: Application of Kalman filter for sensor fusion. In Proceedings of the 2020 IEEE Sensors Applications Symposium (2020).
- 38.Li, C., Fahmy, A. & Sienz, J. An augmented reality based human-robot interaction interface using Kalman filter sensor fusion. Sensors19, 4586. 10.3390/s19204586 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 39.Puranik, A. S. & Swaroop Swamy, K. M. Sensor fusion using Kalman filter in autonomous vehicles. Int. Res. J. Eng. Technol.11(3), 398 (2024). [Google Scholar]
- 40.Tinoco, V. et al. A review of advanced controller methodologies for robotic manipulators. Int. J. Dyn. Control10.1007/s40435-024-01533-1 (2025). [Google Scholar]
- 41.Luan, F., Na, J., Huang, Y. & Gao, G. Adaptive neural network control for robotic manipulators with guaranteed finite-time convergence. Neurocomputing337, 153–164. 10.1016/j.neucom.2019.01.063 (2019). [Google Scholar]
- 42.Khan, A., Abdeljawad, T., Abdel-Aty, M. & Almutairi, D. K. Digital analysis of discrete fractional order cancer model by artificial intelligence. Alex. Eng. J.118, 115–124. 10.1016/j.aej.2025.01.036 (2025). [Google Scholar]
- 43.Khan, A., Abdeljawad, T., Gómez-Aguilar, J. F. & Pérez-Careta, E. Mathematical study of two-strain epidemic model with crossover behavior. Math. Methods Appl. Sci.10.1002/mma.11111 (2025). [Google Scholar]
- 44.Yue, P., Xu, B. & Zhang, M. An improved nonlinear robust control approach for robotic manipulators with PSO-based global optimization strategy. Sci. Rep.14, 21356. 10.1038/s41598-024-72156-x (2024). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45.Khan, M., Khan, S., Palade, V., Mehmood, I. & De Albuquerque, V. H. C. Edge intelligence-assisted smoke detection in foggy surveillance environments. IEEE Trans. Ind. Inf.16(2), 1067–1075. 10.1109/TII.2019.2915592 (2020). [Google Scholar]
- 46.Hafeez, T., Xu, L. & McArdle, G. Edge intelligence for data handling and predictive maintenance in IIOT. IEEE Access9, 49355–49371. 10.1109/ACCESS.2021.3069137 (2021). [Google Scholar]
- 47.Dierkes, E., Meerpohl, C., Flaßkamp, K. & Büskens, C. An adaptive controller for robotic manipulators with unknown kinematics and dynamics. IFAC PapersOnLine54(17), 143–148. 10.1016/j.ifacol.2021.10.342 (2021). [Google Scholar]
- 48.Hsia, T. Adaptive control of robot manipulators—A review. In IEEE International Conference on Robotics and Automation. 10.1109/ROBOT.1991.1087696 (1991).
- 49.Al-Ashoor, R. A., Khorasani, K., Patel, R. V. & Al-Khalili, A. J. Robust adaptive controller design for flexible joint manipulators. Automatica28(2), 327–340. 10.1016/0005-1098(92)90003-O (1992). [Google Scholar]
- 50.Gilbert, E. G., Johnson, D. W. & Keerthi, S. S. Adaptive computed torque control for rigid link manipulations. IEEE J. Robot. Autom.4(2), 207–215. 10.1109/56.2083 (1988). [Google Scholar]
- 51.Piardi, L. F., Oliveira, A. S., Costa, P. & Leitao, P. Collaborative fault tolerance for cyber–physical systems: The detection stage. Comput. Ind.157, 104181. 10.1016/j.compind.2025.104181 (2025). [Google Scholar]
- 52.Qviro. 2025 key trends in collaborative robots. Qviro Blog. https://qviro.com/blog/2025-key-trends-cobots/ (2024).
- 53.Bui, D. N. & Phung, M. D. Radial basis function neural networks for formation control of unmanned aerial vehicles. Robotica42(6), 1842–1860. 10.1017/S0263574724000559 (2024). [Google Scholar]
- 54.Mounish, V. What are the radial basis functions neural networks? Analytics Vidhya Blog. https://www.analyticsvidhya.com/blog/2024/07/radial-basis-functions-neural-networks/ (2024).
- 55.Zhang, Y. & Zhao, L. Adaptive control and state error prediction of flexible manipulators using radial basis function neural network and dynamic surface control method. PLoS ONE20, e0318601. 10.1371/journal.pone.0318601 (2025). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 56.Chen, J. & Ran, X. Deep learning with edge computing: A review. Proc. IEEE107(8), 1655–1674. 10.1109/JPROC.2019.2921977 (2019). [Google Scholar]
- 57.Mohammadi, M., Al-Fuqaha, A., Sorour, S. & Guizani, M. Deep learning for IoT big data and streaming analytics: A survey. IEEE Commun. Surv. Tutor.20(4), 2923–2960. 10.1109/COMST.2018.2844341 (2018). [Google Scholar]
- 58.Liu, J., Jian, F. & Liu, J. Review on motion planning of robotic manipulator in dynamic environments. J. Sens.2024, 5969512. 10.1155/2024/5969512 (2024). [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Data Availability Statement
The datasets generated and analyzed during the current study are available from the corresponding author upon reasonable request. Experimental data including sensor measurements, control parameters, and performance metrics are stored in accordance with institutional data management policies. Software code and algorithm implementations are available under restricted access due to potential commercial applications. Researchers interested in accessing specific datasets for validation or comparative studies should contact the corresponding author with detailed research proposals.



























