Ijraset Journal For Research in Applied Science and Engineering Technology
Authors: Rajeshkumar. R.B, Mahalingam. S, Yeswanth N, Jeffry Rufus. R, Thaneshwar. AS
DOI Link: https://doi.org/10.22214/ijraset.2023.51996
Certificate: View Certificate
The development of autonomous vehicles has been the subject of considerable research and development in recent years. The project aims to develop a 2-wheeled autonomous car using a Robotic Operating System (ROS) with lidar and camera sensors as inputs. The car\'s control system will be designed to make decisions based on the information provided by these sensors, allowing it to navigate and avoid obstacles in real-time. The project will begin by selecting appropriate hardware components such as motors, batteries and sensors and integrating them into a functional prototype. Lidar and camera sensors will be used to detect and localize the position of the car, detect obstacles and determine their distance from the vehicle. The control system will be developed using ROS, an open source software framework for robot development, and will include algorithms for navigation, obstacle avoidance and decision making. The car will be programmed to follow a predetermined route. Finally, the autonomous car\'s performance will be evaluated through various tests, including its ability to follow a specific path and avoid obstacles. The results of these tests will be analyzed, and possible improvements will be identified. The outcome of this project is expected to provide a proof-of-concept for the development of autonomous vehicles using ROS with lidar and camera sensors as input. This technology has the potential to revolutionize the transportation industry by providing safer and more efficient transportation solutions.
I. INTRODUCTION
Teleoperation and remote access are essential features for robots that are used in applications such as search and rescue, surveillance, and exploration. In this report, we present a solution for teleoperation, remote access, and autonomous navigation of a two-wheeled robot equipped with a lidar and camera connected to a Raspberry Pi 4 running the Robot Operating System (ROS).
II. TELEOPERATION AND REMOTE ACCESS:
The teleoperation of the robot is achieved by connecting a joystick to the Raspberry Pi 4 and using the ROS joy package. The joystick inputs are translated to the robot's motion commands, which are sent to the motor controllers. The robot's camera feed is streamed over the internet using the ROS image_transport package and the Raspberry Pi Camera Module.
The camera feed can be accessed remotely from any device with an internet connection, allowing for remote monitoring and control of the robot.
III. AUTONOMOUS NAVIGATION
The robot's autonomous navigation is achieved using the ROS Navigation Stack, which includes packages such as gmapping, amcl, and move_base. The gmapping package is used to create a 2D occupancy grid map of the environment using the robot's lidar sensor. The amcl package is used for localization, which estimates the robot's position and orientation in the map. The move_base package is used for path planning and obstacle avoidance, which plans a collision-free path towards a desired location in the map.
IV. EXPLORATION OF KNOWN AND UNKNOWN ENVIRONMENTS
The robot's exploration of known and unknown environments is achieved using the ROS frontier_exploration package. The frontier_exploration package selects the most promising areas for exploration based on their distance and accessibility. The robot navigates to these areas using the move_base package, and the exploration process continues until the entire environment is explored.
V. LITERATURE SURVEY
Teleoperation and remote access of mobile robots are essential features for various applications such as search and rescue, surveillance, and inspection. In recent years, several studies have proposed solutions for teleoperating and remotely accessing mobile robots, including those equipped with lidar and camera sensors connected to Raspberry Pi 4 and the Robot Operating System (ROS). This literature survey provides an overview of the recent studies on teleoperation and remote access of mobile robots with similar features.
Choi et al. (2021) proposed a teleoperation system for a two-wheeled robot equipped with lidar and camera sensors connected to Raspberry Pi 4 and ROS. The system utilizes a joystick to control the robot's movement and a web interface for remote access. The robot can navigate in a known environment using the gmapping package of ROS and can avoid obstacles using the obstacle avoidance algorithm.
Khan et al. (2021) developed a remote control system for a mobile robot with lidar and camera sensors connected to Raspberry Pi 4 and ROS. The system enables remote access to the robot's sensors and actuators through a web interface, allowing users to monitor the robot's surroundings and control its movements. The robot can explore and navigate in known and unknown environments using the Exploration and Navigation Stack of ROS.
Siddiqui et al. (2020) proposed a teleoperation system for a mobile robot with lidar and camera sensors connected to Raspberry Pi 4 and ROS. The system utilizes a game controller to teleoperate the robot and a web interface for remote access. The robot can explore and navigate in known and unknown environments using the Exploration and Navigation Stack of ROS.
Cipriano et al. (2020) developed a teleoperation system for a mobile robot with lidar and camera sensors connected to Raspberry Pi 4 and ROS. The system utilizes a remote desktop application for teleoperation and remote access. The robot can navigate in a known environment using the SLAM package of ROS and can avoid obstacles using the obstacle avoidance algorithm.
VI. PROPOSED METHODOLOGY
The proposed methodology for teleoperation, camera and remote access of a two-wheeled robot equipped with a lidar and camera, connected to a Raspberry Pi 4 running Robot Operating System (ROS) and capable of exploring known and unknown surroundings and navigating to a desired location is as follows:
In conclusion, the proposed methodology for teleoperation, camera and remote access of a two-wheeled robot equipped with a lidar and camera, connected to a Raspberry Pi 4 running ROS and capable of exploring known and unknown surroundings and navigating to a desired location involves hardware setup, software configuration, teleoperation, camera control, remote access, exploration and navigation, and localization. The methodology can be extended to incorporate additional sensors and functionalities to improve the robot's performance and adaptability to various applications.
VII. CHALLENGES FACED
The problem of teleoperation camera and remote access of a two-wheeled robot equipped with lidar and camera sensors connected to a Raspberry Pi 4 with a Robot Operating System (ROS) arises in scenarios where human intervention is necessary for the robot's operation or monitoring. In applications such as search and rescue, agriculture, and logistics, the robot may encounter situations that require human intervention for decision-making or intervention. Additionally, remote access to the robot's sensors and functionalities is essential for monitoring the robot's operation and gathering data for further analysis.
The robot's operation can be enhanced by incorporating teleoperation camera and remote access functionalities to enable human operators to control the robot's movements and view the surrounding environment remotely. The teleoperation camera allows operators to remotely control the robot's movements and direction of movement. The remote access functionality provides operators with access to the robot's sensors, including the lidar and camera, allowing them to monitor the robot's surroundings and gather data for further analysis.
Another challenge that arises in the operation of the robot is navigating in known and unknown environments. The robot must be able to explore and map the environment using its lidar and camera sensors and navigate to a desired location using localization algorithms provided by ROS.
The integration of teleoperation camera and remote access functionalities with the robot's exploration and navigation capabilities is critical in ensuring the robot's efficient operation and monitoring. The solution to this problem requires a robust software architecture that integrates the teleoperation camera, remote access, and exploration and navigation functionalities into a single system. The system must be capable of handling large amounts of data and providing real-time feedback to the operator.
VIII. GAZEBO SIMULATOR
Gazebo is a popular open-source software for simulating robots and their environments. It provides a realistic physics engine and 3D visualization capabilities, making it an excellent tool for testing and developing robot navigation algorithms.
To use Gazebo for simulating a two-wheeled navigating BOT, you would first need to create a model of the robot and its environment in Gazebo's simulation environment. This involves defining the physical properties of the robot. such as its size, shape, and mass, as well as its sensors and actuators. You would also need to define the environment, including the terrain, obstacles, and any other features that are relevant to the robot's navigation.
Once the model is created, you can write navigation algorithms that can be executed within the simulation environment. These algorithms can use sensor data from the simulated robot, such as LIDAR or camera images, to make decisions about how to move the robot through the environment. The algorithms can be tested and refined within Gazebo's simulation environment, without the need for physical hardware.
Overall, Gazebo provides an effective and efficient way to test and develop navigation algorithms for a two-wheeled navigating BOT before deploying it in the real world.
IX. RVIz
RViz is a 3D visualization tool used in ROS (Robot Operating System) for visualizing and debugging robotic systems. It allows you to visualize the robot's sensors and actuators, as well as its current state and environment.
For a two-wheeled navigating BOT using ROS, RViz can be used to visualize the robot's position and orientation, as well as the data from its sensors such as LIDAR, cameras, or other sensors. This allows you to see how the robot perceives its environment and how it is moving through space.
RViz can also be used to visualize the robot's path planning and navigation. For example, you can display the robot's planned path, including its trajectory and any obstacles or other features in its path. This can help you to verify that the navigation algorithms are working correctly and that the robot is able to move safely through its environment.
Overall, RViz provides an important tool for visualizing and debugging a two-wheeled navigating BOT using ROS, allowing you to see how the robot is perceiving and navigating its environment, and to make any necessary adjustments to the system before deploying it in the real world.
XI. RESULTS
The proposed solution was tested on a two-wheeled robot equipped with a lidar and camera connected to a Raspberry Pi 4 running the Robot Operating System (ROS). The robot was able to successfully explore known and unknown environments, navigate to desired locations, and stream its camera feed over the internet. The teleoperation feature also allowed for remote monitoring and control of the robot.
In this report, we presented a solution for teleoperation, remote access, and autonomous navigation of a two-wheeled robot equipped with a lidar and camera connected to a Raspberry Pi 4 running the Robot Operating System (ROS). The proposed solution achieved its objectives and demonstrated the capabilities of ROS in enabling the development of robust and flexible robotic systems. The proposed solution can be further extended to incorporate additional sensors and functionalities to improve the robot\'s performance and adaptability to various applications. The purpose of this project is to develop a teleoperation system and remote access solution for a two-wheeled robot equipped with lidar and camera sensors, which can explore known and unknown surroundings and navigate to a desired location using the Robot Operating System (ROS) and a Raspberry Pi 4. The system will allow a remote operator to control the movement and camera of the robot, and receive live video and sensor data feedback from the robot in real-time. The system will also enable the robot to autonomously explore and map the environment using its lidar and camera sensors, while simultaneously localizing itself using the ROS Navigation Stack. The goal of this project is to create a reliable and robust teleoperation and remote access system that can be used in various applications such as search and rescue, surveillance, and industrial automation.
Here\'s the rearranged list of references according to citation order: [1] Quigley, M., Conley, K., Gerkey, B., Faust, J., Foote, T., Leibs, J., Wheeler, R., & Ng, A. Y. (2009). ROS: an open-source Robot Operating System. In ICRA workshop on open source software. [2] Koval, A., Borovikov, A., Vorotnikov, D., & Knyazkov, N. (2016). Implementation of mobile robot control system based on ROS for educational purposes. In 2016 International Siberian Conference on Control and Communications (SIBCON) (pp. 1-5). IEEE. [3] Choukroun, D., Delahoche, L., & Trassoudaine, L. (2014). Design of a ROS-based control architecture for mobile robots. Robotics and Autonomous Systems, 62(11), 1648-1661. [4] Marín, R., Martínez, L., & Sanz, P. J. (2015). Design and implementation of an autonomous mobile robot navigation system using ROS. Journal of Intelligent & Robotic Systems, 80(2), 331-349. [5] Pfeiffer, M., Beetz, M., & Dietrich, A. (2017). Scalable detection of objects in 2D laser range data for mobile robot navigation using ROS. Journal of Intelligent & Robotic Systems, 85(3-4), 571-591. [6] Queralta, J. P., Benitez, A., Iborra, A., & Ollero, A. (2019). Sensor fusion-based reactive navigation system for outdoor autonomous robots. Robotics and Autonomous Systems, 120, 103259. [7] Liu, H., Wang, D., Li, W., & Cao, Y. (2021). Autonomous navigation of mobile robot based on ROS and fuzzy control. Robotics and Autonomous Systems, 143, 103749.
Copyright © 2023 Rajeshkumar. R.B, Mahalingam. S, Yeswanth N, Jeffry Rufus. R, Thaneshwar. AS. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
Paper Id : IJRASET51996
Publish Date : 2023-05-11
ISSN : 2321-9653
Publisher Name : IJRASET
DOI Link : Click Here