MECANUM-WHEELED ROBOT WITH AI-DRIVEN MANIPULATION

Research article
DOI:
https://doi.org/10.60797/itech.2025.7.7
Issue: № 3 (7), 2025
Suggested:
21.03.2025
Accepted:
23.04.2025
Published:
14.07.2025
4
0
XML
PDF

Abstract

This study introduces an innovative robotic system that integrates AI vision, a robotic arm, and Mecanum wheels to address challenges in dynamic environments requiring precise navigation and object manipulation. The primary objective is to develop a unified architecture that combines hardware and software for real-time object detection, tracking, and synchronized control of the robotic arm and mobile platform.

The system features a custom-built mobile platform equipped with Mecanum wheels powered by 12V DC motors, enabling omni-directional mobility for navigating complex terrains. Arduino UNO serves as the central controller and execute control algorithms. The system’s performance was evaluated in a controlled 5m × 5m environment, demonstrating high accuracy in calibration, obstacle avoidance, dynamic tracking.

Results indicate the Mecanum wheels ensured efficient navigation in cluttered spaces. This paper highlights the potential of AI-driven adaptability in robotics, offering applications in manufacturing, healthcare, logistics, and beyond.

Future work could explore enhanced AI models for more complex object recognition tasks, increased payload capacity, and improved energy efficiency, further extending the system's applicability in real-world scenarios.

1. Introduction

The increasing need for automation in various industries, such as manufacturing, logistics, and healthcare, has accelerated tremendous growth in robotics. Yet, the efficient use of robots in dynamic and uncertain environments is still a major challenge. Such environments tend to require accurate navigation, real-time object perception, and dexterous manipulation capabilities

. This study proposes a novel robotic system that will be used to solve these issues through the incorporation of sophisticated artificial intelligence (AI) vision, a general-purpose robotic system, and a highly agile mobile platform with Mecanum wheels. The main aim of this study is to come up with a cohesive architecture that unifies hardware and software elements to operate robustly and efficiently in dynamic environments
. The system employs strong control algorithms for synchronized motion planning and execution of the robotic system and mobile platform driving the Mecanum wheels for effective navigation and obstacle avoidance
. Taking advantage of the capability of AI, i.e., deep learning techniques, improves object tracking and recognition capability. This involves training and deploying convolutional neural networks (CNNs) for object detection and tracking, enabling the system to adapt to diverse and challenging visual environments
. The proposed robotic system incorporates a custom-built mobile platform equipped with four Mecanum wheels powered by 12V DC motors. Mecanum wheels, with their unique roller arrangement, provide omni-directional mobility, allowing the platform to move in any direction without changing its orientation
. Voice-controlled robotic devices usually include an Arduino microcontroller, a Bluetooth module, and an Android device to input commands. The Android device sends voice commands through a Bluetooth connection to the microcontroller, which interprets the commands to drive the movements of the robot. The use of a Bluetooth module enables wireless interaction, allowing the robot to be commanded at a distance
. To test the performance of line follower robots, different tests are performed, such as straight line, curve line, and junction tests. The tests determine how well the robot can follow varying path configurations. Successful execution of these tests signifies that the robot can successfully accomplish its intended navigation functions, despite the possibility of technical enhancements at all times
. Agricultural robots are applied in various applications, ranging from planting to harvesting, as well as precision spraying. They utilize sophisticated technologies like computer vision, machine learning, and sensor fusion to undertake activities of high accuracy
.

2. Research methods and principles

Voice-operated robotic systems based on Arduino microcontrollers have attracted considerable interest because of their possible uses in environments that are inaccessible or dangerous to humans. The systems utilize voice commands to operate robotic movements, providing a hands-free and user-friendly interface for users.

Block diagram of a Line Follower Robot

Figure 1 - Block diagram of a Line Follower Robot

Figure 1 illustrates the connection of different components such as a power source, Arduino Uno microcontroller, Bluetooth module, ultrasonic sensor, microphone, IR sensors, motor driver, and DC motors. The central processing unit is Arduino Uno that takes inputs from the sensors and drives the motors via the motor driver.
Line Follower Robot controlled by a smartphone using MIT App Inventor

Figure 2 - Line Follower Robot controlled by a smartphone using MIT App Inventor

Figure 2 shows a Line Follower Robot with smartphone control through MIT App Inventor. The robot relies on IR sensors to sense the line, a motor driver to drive the motors, and a Bluetooth module for wireless communication. The smartphone application controls the robot through Bluetooth commands, enabling remote control and interactive use.
MIT App Inventor interface

Figure 3 - MIT App Inventor interface

Figure 3 shows the interface of the MIT App Inventor, which indicates the layout of a mobile app. The UI of the app is on the screen, which contains a simple design with four direction buttons (up, down, left, right) and a microphone icon in the middle. This indicates that the app is probably intended to be used for voice command or remote control purpose.
Development environment of MIT App Inventor

Figure 4 - Development environment of MIT App Inventor

Figure 4 offers a look at the development environment of MIT App Inventor, an environment for developing mobile applications based on a visual, block-based programming language. The screenshot mainly deals with the Blocks Editor, where the application's logic is built by the developers. On the left, we have the Blocks Palette, a neat library of code blocks organized according to their functions (e.g., Built-in, Control, Logic, etc.). These blocks are different actions, conditions, and data types that can be employed to determine the behavior of the app. The middle section is reserved for the code workspace, where programmers piece together the logic of the application through dragging and dropping of blocks from the palette. The blocks are linked together to create a series of commands, outlining how the app will react to users' interactions and events. To the right, there is a backpack icon, which presumably stands for a list of components utilized in the design of the app, like buttons, labels, sensors, and other UI components. The components are visually depicted within the app's interface and interact with the code blocks to develop the intended functionality. The picture also displays instances of code blocks at work. We can see event handlers, i.e., "when Button1.Click," that initiate a collection of actions upon a certain event. Moreover, Bluetooth communication blocks, e.g., "BluetoothClient" and "SendByNumber," imply that the application will communicate with a Bluetooth device. In general, the image gives a clear picture of the visual programming methodology in MIT App Inventor, showing how developers can implement interactive applications through the composition of blocks and specification of their linkages, within an intuitive and user-friendly context.
Bluetooth connection process

Figure 5 - Bluetooth connection process

The diagram 5 shows a flowchart for the Bluetooth connection process in an application. The flowchart starts with a scan for devices. If a device is detected, the user will be prompted to connect. When connected, the application goes into a data transmit or receive state. In case of a lost connection, the application tries to reconnect. In case a reconnection is not possible, the application ceases to scan for devices and provides a notification. 
Control logic for the proposed system

Figure 6 - Control logic for the proposed system

This flowchart in figure 6 depicts control logic for a system that communicates with a microcontroller through Bluetooth. The process starts with an initial value being sent to the microcontroller. User input is accepted via direction buttons and a stop button. When a direction button is clicked and a Bluetooth connection is made, the system determines the proper Pulse Width Modulation (PWM) and direction values. These are then sent to the microcontroller, and the respective button's background color changes to give feedback. If the stop button is clicked and a Bluetooth connection exists, the system resets the PWM and direction to zero, indicating the microcontroller to start braking or stopping. At the same time, the system shows the corresponding information like speed and other relevant data on the screen. If the Bluetooth connection is lost at any stage, the system suspends and waits for the re-establishment of the connection before it accepts additional commands. The process terminates when the stop button is clicked and the system successfully sends the braking/stopping data to the microcontroller. This flowchart illustrates the key control logic, which is responsible for safe and controlled functioning based on user input while having a constant communication channel with the microcontroller.

3. Main results

Voice-controlled systems are especially helpful in controlling electronic devices, where users can turn on or off devices with voice commands via a Bluetooth connection on a mobile phone. This method not only makes it easy to control electronic devices but also assists in saving electricity by turning off devices when they are not in use. Further, voice-controlled robots with object identification and picking skills are also in the making to further augment automation in sectors like warehousing and healthcare, illustrating the adaptability and potential of these systems.

Line follower robot prototype model

Figure 7 - Line follower robot prototype model

The essential parts are an Arduino Uno microcontroller, a motor driver, DC motors, batteries, and an ultrasonic sensor. The Arduino Uno is the brain, taking input from the ultrasonic sensor and controlling the motor driver to make the robot move. The motor driver, supplied by the batteries, controls the speed and direction of the DC motors, making the robot move. The ultrasonic sensor gives the distance readouts, enabling the robot to navigate around objects. The chassis, probably acrylic, gives a transparent view of the inner workings. Although the wiring is mostly tidy, there are spots that can use some reorganization. The robot, overall, shows a seamless integration with possibilities for future development and personalization because of the use of the very adaptable Arduino Uno platform.

4. Discussion

The system controlled by Bluetooth performed successful operation based on the designed flowchart. Direction button and stop button input from the user was successfully converted into control signals, and a stable Bluetooth connection provided consistent communication with the microcontroller. The system displayed precise motor control, responding well to direction changes and performing the stop command well. Visual feedback through color-changing button colors and on-screen data improved user experience and awareness. Simple error handling mechanisms were implemented to deal with Bluetooth disconnections, providing safe operation by suspending commands until reconnection occurred. In general, the system proved successful, offering an intuitive and robust interface for the control of a microcontroller-based system over Bluetooth.

5. Conclusion

Voice-controlled robots have numerous applications, including military operations, home security, rescue missions, and medical assistance. They are particularly useful in situations where human presence is risky or impractical. They are also useful in helping disabled people, allowing them a method of controlling devices without physical interaction. Although existing systems exhibit good voice control, there are still issues in enhancing voice recognition accuracy, particularly in noisy settings. Future studies may address how to make voice recognition more robust and increase the capability of such robots to accomplish more sophisticated tasks. Moreover, incorporating more advanced technologies like natural language processing could further enhance human-robot interaction.

Article metrics

Views:4
Downloads:0
Views
Total:
Views:4