-
Notifications
You must be signed in to change notification settings - Fork 494
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
* tyde CI * getting a corrcet init location of arm * tidy CI * sensor clean up so that the size is the same * setup social nav reward and yaml * bounding box update * clean up human bbox detector * tidy * tidy * add stereo pari camera * make the size of HumanoidDetectorSensor consistent with the senser obs * fix nan issue * fix init_distributed issue * tidy * tidy CI and add comments * add more comments * tidy CI * add InitialGpsCompassSensor sensor * refine social nav reward * add exploration reward * increase the video size * polsih the PR * tidy * reduce the third_rgb_sensor size --------- Co-authored-by: Jimmy Yang <[email protected]>
- Loading branch information
1 parent
ecddd75
commit 461ed57
Showing
11 changed files
with
417 additions
and
75 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -4,7 +4,7 @@ defaults: | |
- /habitat: habitat_config_base | ||
- /habitat/task: task_config_base | ||
# Config agent sensors | ||
- /habitat/simulator/[email protected]_0: rgbd_head_rgbdp_arm_agent | ||
- /habitat/simulator/[email protected]_0: spot_agent | ||
- /habitat/simulator/[email protected]_0: spot | ||
- /habitat/simulator/[email protected]_1: rgbd_head_agent | ||
- /habitat/simulator/[email protected]_1: human | ||
|
@@ -19,17 +19,14 @@ defaults: | |
- /habitat/task/[email protected]_1_oracle_nav_randcoord_action: oracle_nav_action | ||
# Config agent sensor | ||
- /habitat/task/measurements: | ||
- articulated_agent_force | ||
- articulated_agent_colls | ||
- num_steps | ||
- did_agents_collide | ||
- num_agents_collide | ||
- composite_stage_goals | ||
- pddl_subgoal_reward | ||
- rearrange_cooperate_reward | ||
- articulated_agent_force | ||
- force_terminate | ||
- social_nav_reward | ||
- object_to_goal_distance | ||
- articulated_agent_colls | ||
- rot_dist_to_goal | ||
- dist_to_goal | ||
- nav_to_pos_succ | ||
|
@@ -50,9 +47,17 @@ defaults: | |
- humanoid_detector_sensor | ||
- has_finished_oracle_nav | ||
- /habitat/task/[email protected]_sensors.agent_0_goal_to_agent_gps_compass: nav_goal_sensor | ||
- /habitat/task/[email protected]_sensors.agent_0_spot_head_stereo_depth_sensor: spot_head_stereo_depth_sensor | ||
- /habitat/task/[email protected]_sensors.agent_1_agents_within_threshold: agents_within_threshold | ||
- _self_ | ||
|
||
# This yaml is designed specifically for learning a social nav policy for the Boston Dynamics Spot robot. | ||
# The major changes compared to the base yaml are the following: | ||
# (1) obs_keys: we ensure these observations can be obtained from the real robot (Spot) | ||
# (2) social_nav_reward: social nav reward considers collisions based on a kinematic simulation | ||
# (3) actions: Spot controls the base linear and angualr velocity | ||
# (4) simulator: we simulate the environment via the kinematic mode to facilitate sim2real transfer | ||
|
||
habitat: | ||
task: | ||
lab_sensors: | ||
|
@@ -62,6 +67,11 @@ habitat: | |
agent_idx: 1 | ||
agent_0_goal_to_agent_gps_compass: | ||
goal_is_human: True | ||
# Defien the human detector | ||
humanoid_detector_sensor: | ||
# If the human detector function is image or binary flag | ||
return_image: False | ||
is_return_image_bbox: False | ||
type: RearrangePddlSocialNavTask-v0 | ||
min_start_distance: 4.0 | ||
reward_measure: social_nav_reward | ||
|
@@ -70,20 +80,28 @@ habitat: | |
slack_reward: -0.01 | ||
end_on_success: False | ||
task_spec_base_path: benchmark/multi_agent/ | ||
task_spec: pddl/multi_agent_tidy_house | ||
task_spec: pddl/multi_agent_social_nav | ||
pddl_domain_def: fp | ||
min_distance_start_agents: 5.0 | ||
constraint_violation_ends_episode: False | ||
constraint_violation_drops_object: True | ||
|
||
measurements: | ||
rearrange_cooperate_reward: | ||
stage_sparse_reward: 5.0 | ||
end_on_collide: True | ||
collide_penalty: 0.5 | ||
# We do kinemantic simulation here, so no dynamic force | ||
force_terminate: | ||
max_accum_force: -1 | ||
max_instant_force: -1 | ||
# Define the social nav reward | ||
social_nav_reward: | ||
facing_human_reward: 0.01 | ||
count_coll_pen: 0.01 | ||
max_count_colls: 1000 | ||
count_coll_end_pen: 1.0 | ||
# Define the social nav seek success | ||
social_nav_seek_success: | ||
must_look_at_targ: False | ||
following_step_succ_threshold: 500 | ||
need_to_face_human: True | ||
|
||
actions: | ||
agent_0_base_velocity: | ||
|
@@ -102,6 +120,7 @@ habitat: | |
ang_speed: 5.0 | ||
enable_rotation_check_for_dyn_slide: False | ||
|
||
# Human random walking action | ||
agent_1_oracle_nav_action: | ||
type: OracleNavAction | ||
motion_control: human_joints | ||
|
@@ -121,7 +140,7 @@ habitat: | |
gym: | ||
obs_keys: | ||
- agent_0_articulated_agent_arm_depth | ||
- agent_0_goal_to_agent_gps_compass | ||
- agent_0_spot_head_stereo_depth_sensor | ||
- agent_0_humanoid_detector_sensor | ||
- agent_1_head_depth | ||
- agent_1_relative_resting_position | ||
|
@@ -150,27 +169,8 @@ habitat: | |
agents_order: | ||
- agent_0 | ||
- agent_1 | ||
agents: | ||
agent_0: | ||
radius: 0.25 | ||
height: 1.41 | ||
articulated_agent_urdf: data/robots/hab_spot_arm/urdf/hab_spot_arm.urdf | ||
articulated_agent_type: "SpotRobot" | ||
sim_sensors: | ||
arm_rgb_sensor: | ||
height: 480 | ||
width: 640 | ||
hfov: 47 | ||
arm_depth_sensor: | ||
height: 224 | ||
width: 171 | ||
hfov: 55 | ||
arm_panoptic_sensor: | ||
height: 224 | ||
width: 171 | ||
hfov: 55 | ||
joint_start_noise: 0.0 | ||
|
||
# We use the kinematic mode to train the real world social nav policy | ||
kinematic_mode: True | ||
ac_freq_ratio: 1 | ||
step_physics: False | ||
|
58 changes: 58 additions & 0 deletions
58
habitat-lab/habitat/config/benchmark/multi_agent/pddl/multi_agent_social_nav.yaml
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,58 @@ | ||
# Version of tidy house that only has two target objects. | ||
objects: | ||
- name: any_targets|0 | ||
expr_type: movable_entity_type | ||
- name: TARGET_any_targets|0 | ||
expr_type: goal_entity_type | ||
- name: any_targets|1 | ||
expr_type: movable_entity_type | ||
- name: TARGET_any_targets|1 | ||
expr_type: goal_entity_type | ||
- name: robot_0 | ||
expr_type: robot_entity_type | ||
- name: robot_1 | ||
expr_type: robot_entity_type | ||
|
||
goal: | ||
expr_type: AND | ||
sub_exprs: | ||
- at(any_targets|0,TARGET_any_targets|0) | ||
- at(any_targets|1,TARGET_any_targets|1) | ||
- not_holding(robot_0) | ||
- not_holding(robot_1) | ||
|
||
stage_goals: | ||
stage_1_1: | ||
expr_type: AND | ||
sub_exprs: | ||
- quantifier: EXISTS | ||
inputs: | ||
- name: robot | ||
expr_type: robot_entity_type | ||
expr_type: AND | ||
sub_exprs: | ||
- holding(any_targets|0, robot) | ||
stage_1_2: | ||
expr_type: AND | ||
sub_exprs: | ||
- at(any_targets|0,TARGET_any_targets|0) | ||
stage_2_1: | ||
expr_type: AND | ||
sub_exprs: | ||
- quantifier: EXISTS | ||
inputs: | ||
- name: robot | ||
expr_type: robot_entity_type | ||
expr_type: AND | ||
sub_exprs: | ||
- holding(any_targets|1, robot) | ||
stage_2_2: | ||
expr_type: AND | ||
sub_exprs: | ||
- at(any_targets|1,TARGET_any_targets|1) | ||
|
||
solution: | ||
- nav_to_goal(any_targets|1, robot_1) | ||
- pick(any_targets|1, robot_1) | ||
- nav_to_obj(TARGET_any_targets|1, robot_1) | ||
- place(any_targets|1,TARGET_any_targets|1, robot_1) |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Oops, something went wrong.