Giter Site home page Giter Site logo

quantum-entropy / elevation_mapping Goto Github PK

View Code? Open in Web Editor NEW

This project forked from ewenwan/elevation_mapping

0.0 0.0 0.0 1.37 MB

高程制图-粗糙地形导航

Home Page: https://cloud.tencent.com/developer/news/132290

License: BSD 3-Clause "New" or "Revised" License

CMake 3.44% C++ 95.31% Python 1.25%

elevation_mapping's Introduction

Robot-Centric Elevation Mapping

Overview

This is a ROS package developed for elevation mapping with a mobile robot. The software is designed for (local) navigation tasks with robots which are equipped with a pose estimation (e.g. IMU & odometry) and a distance sensor (e.g. structured light (Kinect, RealSense), laser range sensor, stereo camera). The provided elevation map is limited around the robot and reflects the pose uncertainty that is aggregated through the motion of the robot (robot-centric mapping). This method is developed to explicitly handle drift of the robot pose estimation.

The Robot-Centric Elevation Mapping packages have been tested under ROS Kinetic and Ubuntu 16.04. This is research code, expect that it changes often and any fitness for a particular purpose is disclaimed.

Author: Péter Fankhauser
Affiliation: ANYbotics
Maintainer: Péter Fankhauser, [email protected]

This projected was initially developed at ETH Zurich (Autonomous Systems Lab & Robotic Systems Lab).

This work is conducted as part of ANYmal Research, a community to advance legged robotics.

Build Status

Elevation Map Example

Videos of the elevation mapping software in use:

Citing

The robot-centric elevation mapping methods used in this software are described in the following paper (available here). If you use this work in an academic context, please cite the following publication(s):

  • P. Fankhauser, M. Bloesch, and M. Hutter, "Probabilistic Terrain Mapping for Mobile Robots with Uncertain Localization", in IEEE Robotics and Automation Letters (RA-L), vol. 3, no. 4, pp. 3019–3026, 2018. (PDF)

      @article{Fankhauser2018ProbabilisticTerrainMapping,
        author = {Fankhauser, P{\'{e}}ter and Bloesch, Michael and Hutter, Marco},
        doi = {10.1109/LRA.2018.2849506},
        title = {Probabilistic Terrain Mapping for Mobile Robots with Uncertain Localization},
        journal = {IEEE Robotics and Automation Letters (RA-L)},
        volume = {3},
        number = {4},
        pages = {3019--3026},
        year = {2018}
      }
    
  • P. Fankhauser, M. Bloesch, C. Gehring, M. Hutter, and R. Siegwart, "Robot-Centric Elevation Mapping with Uncertainty Estimates", in International Conference on Climbing and Walking Robots (CLAWAR), 2014. (PDF)

      @inproceedings{Fankhauser2014RobotCentricElevationMapping,
        author = {Fankhauser, P\'{e}ter and Bloesch, Michael and Gehring, Christian and Hutter, Marco and Siegwart, Roland},
        title = {Robot-Centric Elevation Mapping with Uncertainty Estimates},
        booktitle = {International Conference on Climbing and Walking Robots (CLAWAR)},
        year = {2014}
      }
    

Installation

Dependencies

This software is built on the Robotic Operating System (ROS), which needs to be installed first. Additionally, the Robot-Centric Elevation Mapping depends on following software:

Building

In order to install the Robot-Centric Elevation Mapping, clone the latest version from this repository into your catkin workspace and compile the package using ROS.

cd catkin_workspace/src
git clone https://github.com/anybotics/elevation_mapping.git
cd ../
catkin_make

Unit Tests

Run the unit tests with

catkin_make run_tests_elevation_map_msg run_tests_elevation_mapping

Basic Usage

In order to get the Robot-Centric Elevation Mapping to run with your robot, you will need to adapt a few parameters. It is the easiest if duplicate and adapt all the parameter files that you need to change from the elevation_mapping_demos package (e.g. the simple_demo example). These are specifically the parameter files in config and the launch file from the launch folder.

Nodes

Node: elevation_mapping

This is the main Robot-Centric Elevation Mapping node. It uses the distance sensor measurements and the pose and covariance of the robot to generate an elevation map with variance estimates.

Subscribed Topics

Published Topics

  • elevation_map (grid_map_msg/GridMap)

    The entire (fused) elevation map. It is published periodically (see fused_map_publishing_rate parameter) or after the trigger_fusion service is called.

  • elevation_map_raw (grid_map_msg/GridMap)

    The entire (raw) elevation map before the fusion step.

Services

  • trigger_fusion (std_srvs/Empty)

    Trigger the fusing process for the entire elevation map and publish it. For example, you can trigger the map fusion step from the console with

      rosservice call /elevation_mapping/trigger_fusion
    
  • get_submap (grid_map_msg/GetGridMap)

    Get a fused elevation submap for a requested position and size. For example, you can get the fused elevation submap at position (-0.5, 0.0) and size (0.5, 1.2) and safe it to a text file form the console with

      rosservice call -- /elevation_mapping/get_submap -0.5 0.0 0.5 1.2 []
    
  • clear_map (std_srvs/Empty)

    Initiates clearing of the entire map for resetting purposes. Trigger the map clearing with

      rosservice call /elevation_mapping/clear_map
    

Parameters

  • point_cloud_topic (string, default: "/points")

    The name of the distance measurements topic.

  • robot_pose_topic (string, default: "/robot_state/pose")

    The name of the robot pose and covariance topic.

  • base_frame_id (string, default: "/robot")

    The id of the robot base tf frame.

  • map_frame_id (string, default: "/map")

    The id of the tf frame of the elevation map.

  • track_point_frame_id (string, default: "/robot")

    The elevation map is moved along with the robot following a track point. This is the id of the tf frame in which the track point is defined.

  • track_point_x, track_point_y, track_point_z (double, default: 0.0, 0.0, 0.0)

    The elevation map is moved along with the robot following a track point. This is the position of the track point in the track_point_frame_id.

  • robot_pose_cache_size (int, default: 200, min: 0)

    The size of the robot pose cache.

  • min_update_rate (double, default: 2.0)

    The mininum update rate (in Hz) at which the elevation map is updated either from new measurements or the robot pose estimates.

  • fused_map_publishing_rate (double, default: 1.0)

    The rate for publishing the entire (fused) elevation map.

  • relocate_rate (double, default: 3.0)

    The rate (in Hz) at which the elevation map is checked for relocation following the tracking point.

  • length_in_x, length_in_y (double, default: 1.5, min: 0.0)

    The size (in m) of the elevation map.

  • position_x, position_y (double, default: 0.0)

    The position of the elevation map (center) in the elevation map frame.

  • resolution (double, default: 0.01, min: 0.0)

    The resolution (cell size in m/cell) of the elevation map.

  • min_variance, max_variance (double, default: 9.0e-6, 0.01)

    The minimum and maximum values for the elevation map variance data.

  • mahalanobis_distance_threshold (double, default: 2.5)

    The threshold for the Mahalanobis distance. Decides if measurements are fused with the existing data, overwritten or ignored.

  • multi_height_noise (double, default: 9.0e-7)

    Added noise for cell with multiple height measurements (e.g. walls).

  • min_horizontal_variance, max_horizontal_variance (double, default: pow(resolution / 2.0, 2), 0.5)

    The minimum and maximum values for the elevation map horizontal variance data.

  • enable_visibility_cleanup (bool, default: true)

    Enable visibility cleanup. This runs a separate thread to remove elements in the map based on the visibility constraint.

  • visibility_cleanup_rate (double, default: 1.0)

    The rate (in Hz) at which the visibility constraint is performed.

  • scanning_duration (double, default: 1.0)

    The sensor's scanning duration (in s) which is used for the visibility cleanup. Set this roughly to the duration it takes between two consecutive full scans (e.g. 0.033 for a ToF camera with 30 Hz, or 3 s for a rotating laser scanner). Depending on how dense or sparse your scans are, increase or reduce the scanning duration. Smaller values lead to faster dynamic object removal and bigger values help to reduce faulty map cleanups.

  • sensor_cutoff_min_depth, sensor_cutoff_max_depth (double, default: 0.2, 2.0)

    The minimum and maximum values for the length of the distance sensor measurements. Measurements outside this interval are ignored.

  • sensor_model_normal_factor_a, sensor_model_normal_factor_b, sensor_model_normal_factor_c, sensor_model_lateral_factor (double)

    The data for the sensor noise model.

Bugs & Feature Requests

Please report bugs and request features using the Issue Tracker.

elevation_mapping's People

Contributors

fgiraldez avatar hanneskeller avatar jlack1987 avatar lisler avatar lkotsoni avatar mariushuber avatar mbjelonic avatar mktk1117 avatar pfankhauser avatar remod avatar samuelba avatar

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.