Skip to content

How-to on simulating a robot with V-REP and controlling it with ROS

Notifications You must be signed in to change notification settings

Nurgak/Virtual-Robot-Challenge

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

16 Commits
 
 
 
 
 
 
 
 

Repository files navigation

Virtual Robot Challenge

Image of the virtual robot in V-REP

The goal of the Virtual Robot Challenge (VRC) is to introduce the Robotic Operating System (ROS) and V-REP simulation environment by performing autonomous tasks with a simulated robot.

This is to demonstrate the advantages of development using simulation, before eventually moving on to real robots. The advantages are that the code stays the same when moving from simulation to real robot, testing in simulation is much less time consuming (avoid mechanical failures, batteries, sensors failures/calibration...), cheaper and requires less effort in general.

All software used within the scope of this project is free to use.

Theory

ROS is composed of small stand-alone entities called "nodes". Nodes publish and subscribe to messages called "topics" to talk between each other. One can write nodes in C++ or Python, topics are language agnostic. A good introduction can be found on the ROS official website.

The ROS tutorials are well written and straight forward, it is recommended to follow them as least for installation and workspace setup, this provides a good overview on how ROS works and how it is organized.

Hardware

The robot has two wheels and a caster wheel in the back for locomotion. In the front it has six distance sensors, four pointing forwards and two pointing downwards under the robot, and a linear camera showing an image of 128 pixels wide and 1 pixel height.

The robot is based on the Robopoly demonstration robot and can be physically built, any type of robot can be adapted for the simulation.

Image of the virtual and real robot side by side

Software

The whole installation process has been fully tested on Ubuntu 14.04 with ROS Indigo. One can use VirtualBox if such a computer is not available, however it will slow down the simulation speed.

Installation

  1. Download and install the latest V-REP EDU, define the VREP_ROOT environment variable as the path to V-REP

    export VREP_ROOT=<VREP path>
    
  2. Install ROS Indigo

    $ sudo sh -c 'echo "deb http://packages.ros.org/ros/ubuntu $(lsb_release -sc) main" > /etc/apt/sources.list.d/ros-latest.list'
    $ sudo apt-key adv --keyserver hkp://ha.pool.sks-keyservers.net --recv-key 0xB01FA116
    $ sudo apt-get update
    $ sudo apt-get install ros-indigo-desktop-full
    $ source /opt/ros/indigo/setup.bash
    
  3. Install the catkin tools:

    $ sudo apt-get install python-catkin-tools
    
  4. Create the workspace for ROS in ~/catkin_ws:

    $ mkdir -p ~/catkin_ws/src
    $ cd ~/catkin_ws/src
    $ catkin_init_workspace
    
  5. Copy the ROS interface package contents from V-REP install folder to ROS workspace:

    $ cp -r <VREP path>/programming/ros_packages/v_repExtRosInterface ~/catkin_ws/src
    
  6. Clone this repository to somewhere and copy the contents the ROS package to the workspace:

    $ git clone https://github.com/Nurgak/Virtual-Robot-Challenge.git
    $ cd Virtual-Robot-Challenge
    $ cp -r Software/ROS/vrc ~/catkin_ws/src
    
  7. Build the packages (grab a coffee):

    $ cd ~/catkin_ws
    $ catkin build
    
  8. Copy the ROS interface plugin to the V-REP root folder

    $ cp ~/catkin_ws/devel/lib/libv_repExtRosInterface.so <VREP path>
    
  9. Source the setup file, this needs to be done in every new terminal:

$ cd ~/catkin_ws
$ source devel/setup.bash
  1. Export the ROS_MASTER_URI environment variable
$ export ROS_MASTER_URI=http://localhost:11311

Programming

The code actuating the robot motors and publishing its sensor data is written in Lua language in V-REP. It is very simple and only used to exchange data with the ROS framework, not for actual behavior logic. This is the place to add new code if one wishes to add new sensors to the simulated robot. To view the code open the V-REP scene and on the left side double-click on the script icon next to the robot node.

Image of the robot node with a script icon besides it

The robot behavior is dictated by the Python scripts stored in ~/catkin_ws/src/vrc/src/. The sample.py script shows how to process the data from all the different sensors on the robot.

The launch files are used to start different nodes at the same time, they are stored in ~/catkin_ws/src/vrc/launch/, think of them as shortcuts.

Follow the line

To goal is to follow a line. Very basic approach, however one can always optimize the system to make it faster.

  1. Run ROS core in a terminal

    $ roscore
    
  2. From another terminal run V-REP, always start V-REP after ROS core

    $ sh <VREP path>/vrep.sh
    
  3. In V-REP open the arena.ttt from the /Simulation folder

  4. Click on the play button in V-REP to launch the simulation

    V-REP play button

  5. In a third terminal launch the arena.launch file which will starts some nodes

    $ roslaunch vrc arena.launch
    
  6. Finally in a fourth terminal run the line follower program

    $ rosrun vrc line_follower.py
    

The robot should start following the line marked on the floor.

Image of the robot following a line

Teleoperation

One can remotely operate the robot using the keyboard using the teleop_twist_keyboard package:

$ sudo apt-get install ros-indigo-teleop-twist-keyboard

Stop the line follower script (ctrl+c) and run:

$ rosrun teleop_twist_keyboard teleop_twist_keyboard.py

The terminal must be in focus for the commands to be executed.

Exploration and mapping

ROS can do mapping using the robot sensors and its position. First make sure the slam_gmapping and pointcloud_to_laserscan packages are installed:

$ sudo apt-get install ros-indigo-slam-gmapping
$ sudo apt-get install ros-indigo-pointcloud-to-laserscan

And launch the mapping scene with:

$ roslaunch vrc mapping.launch
$ rosrun vrc explorer.py

The launch file runs the gmapping and the RVIZ visualization nodes. Accessible areas are marked in white and inaccessible areas are marked in black.

The explorer node makes the robot go around the arena and avoid obstacles, whenever it encounters an obstacle it will turn on itself for a random amount of time and continue.

Image of the robot mapping in RVIZ

Physical robot

When the simulation yields satisfying results one can use a real robot that subscribes to and publishes the same ROS topics as the virtual robot, this way the same software running on the computer can be used. The implementation for this can be done in various ways and is left up to the reader.

About

How-to on simulating a robot with V-REP and controlling it with ROS

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published