SLAM and autonomous navigation with ROS + kinect + arduino + android

来源:互联网 发布:梦里花落知多少抄袭 编辑:程序博客网 时间:2024/05/18 03:36

Sung's Blog

CATEGORY: PROJECTS

My Personal Robotic Companion

SLAM and autonomous navigation with ROS + kinect + arduino + android

https://github.com/sungjik/my_personal_robotic_companion

  20150926_21523820150926_215308 20150926_21521320150926_215254 20150926_215135

patrick_the_robot_flow_chart

  1. The Hardware
    – two geared DC motors with integrated encoders (RB30-MM7W-D W/EC 26P 12V)
    {gear ratio: 1/120, rated speed: 49.5 rpm, torque: 4kgf.cm, encoder resolution: 26}
    – Xbox kinect 360
    – Macbook Air 11 inch running Lubuntu 14.04 and ROS indigo
    – Galaxy S6 edge w/ Android
    – Arduino Mega 2560
    – Adafruit Motor Shield v2
    – 6800mAh 3S LiPo battery (and balance charger)
    – DC to DC step up and step down converter (used for connecting the kinect to the battery)
    – LiPo alarm (to prevent completely discharging the battery)
    – Wood planks, metal connectors, nuts, bolts, wires, cables, etc.
  2. Arduino(motor_controller.ino)
    The main loop in motor_controller.ino converts encoder tick counts into the current rpm of both motors and controls the speed of the motors using PID control. The rosserial library conveniently allows the program to communicate with the rest of the ros nodes in the pc via usb. The program subscribes to the cmd_vel topic which sets the desired speed of the motors, and publishes the actual speeds on the rpm topic. I used two interrupt pins in the mega, one for each encoder. The PID control and rpm publishing are all done in the same loop cycle, at a desired rate of 10Hz (every 100 ms). I modifed the Adafruit motorshield v2 library such that setSpeed for DC motors used the full 4096 PWM resolution, i.e. in Adafruit_MotorShield.cpp
    void Adafruit_DCMotor::setSpeed(uint16_t speed) {
    MC->setPWM(PWMpin, speed);
    }

    The uint8_t parameter in the corresponding declaration in Adafruit_MotorShield.h also needs to be modified accordingly. The adafruit motor shield v2 has its own on-board pwm chip with 12 bit resolution, but for some strange reason the dc motor library only used 8 bits.
    References :
    – http://forum.arduino.cc/index.php?topic=8652.0
    – http://ctms.engin.umich.edu/CTMS/index.php?example=Introduction§ion=ControlPID
    – http://playground.arduino.cc/Main/RotaryEncoders
  3. base_controller.cpp
    The base_controller subscribes to the rpm topic from the arduino and converts it into x velocity, theta velocity, xy position, and yaw. It also subscribes to the gyro topic from the android phone and combines the readings with the rpm information to produce a combined yaw. The base_controller then publishes this information in the form of odometry and tf messages.
    Subscribed Topics : rpm(geometry_msgs::Vector3Stamped), gyro(geometry_msgs::Vector3)
    Published Topics : odom(nav_msgs:Odometry)
    Parameters:
    – publish_tf : whether to publish tf or not. set to false if you want to use robot_pose_ekf
    – publish_rate : rate in hz for publishing odom and tf messages
    – linear_scale_positive : amount to scale translational velocity in the positive x direction
    – linear_scale_negative : amount to scale translational velocity in the negative x direction
    – angular_scale_positive : amount to scale rotational velocity counterclockwise
    – angular_scale_negative : amount to scale rotational velocity clockwise
    – alpha : value between 0 and 1, how much to weigh wheel encoder yaw relative to gyro yaw
    References :
    – https://github.com/Arkapravo/turtlebot
    – http://wiki.ros.org/turtlebot
    – http://wiki.ros.org/ros_arduino_bridge
  4. Android(gyro.py)
    I found it necessary to add an IMU to the robot to improve odometry, since odometry from wheel encoders was prone to error from voltage changes, wheel diameter and track width miscalculations, etc. I used an app called HyperImu on the android app market, which publishes gyroscope, accelerometer, magnetometer, and other useful sensor readings onto a UDP stream. I connected my android phone to the PC via usb and set up usb tethering on the android to get an IP address for the UDP stream. Getting the odometry to work with reasonable accuracy was one of the toughest parts of the project. There always existed a discrepancy between what the software thought the robot was doing and what the robot was doing in real life. I at first tried fusing magnetometer, accelerometer, and gyroscope data using a complementary filter and a kalman filter, but in the end I found that a simple weighted average of the gyroscope data and wheel encoder data worked the best. To test whether odometry was reasonably accurate or not, I followed the instructions in http://wiki.ros.org/navigation/Tutorials/Navigation%20Tuning%20Guide.
  5. Kinect
    I followed the instructions in http://wiki.ros.org/kinect/Tutorials/Adding%20a%20Kinect%20to%20an%20iRobot%20Create to connect my 11.1v lipo battery to the kinect. I used an automatic step up/down DC converter instead of a voltage regulator to ensure a steady 12v to the kinect. I then installed the freenect library :
    sudo apt-get install ros-indigo-freenect-stack
    and used the depthimage_to_laserscan ros package to publish a fake laser scan derived from the kinect’s rgbd camera and ir sensors.
    References :
    – https://github.com/turtlebot/turtlebot/blob/indigo/turtlebot_bringup/launch/3dsensor.launch
  6. teleop_twist_keyboard
    The teleop_twist_keyboard ros package takes in keyboard input and publishes cmd_vel messages.
  7. gmapping
    sudo apt-get install ros-indigo-slam-gmapping ros-indigo-gmapping
    The ros gmapping package uses Simultaneous Localization and Mapping(SLAM) to produce a 2D map from laser scan data. I played around with the parameters linearUpdate, angularUpdate, and particles to get a reasonably accurate map. I pretty much followed the directions in http://www.hessmer.org/blog/2011/04/10/2d-slam-with-ros-and-kinect/ and referenced the parameter settings in the turtlebot stack(http://wiki.ros.org/turtlebot).
  8. navigation
    sudo apt-get install ros-indigo-navigation
    I used the amcl, base_local_planner, and costmap_2d packages from the ros navigation stack. I pretty much followed the instructions in http://wiki.ros.org/navigation/Tutorials. One thing I had to do differently was adding map topics for the global and local costmap in rviz instead of a a grid topic. I also modified and used https://code.google.com/p/drh-robotics-ros/source/browse/trunk/ros/ardros/nodes/velocityLogger.py to calculate the acceleration limits of the robot. After hours of tuning amcl, base_local_planner, and costamp_2d parameters, I finally got everything to work.

by Sung Jik Cha

1 0
原创粉丝点击