- To build truely Model Free RL for the current agent we will need following information on which agent learns :
- RGB values of the image showcasing robot from multiple positions at each timestep:
- Need this because we want the robot to learn to corelate from visual features with its own joint positions.
Observation.left_shoulder_rgb: np.ndarray
Observation.left_shoulder_depth: np.ndarray
Observation.left_shoulder_mask: np.ndarray
Observation.right_shoulder_rgb: np.ndarray
Observation.right_shoulder_depth: np.ndarray
Observation.right_shoulder_mask: np.ndarray
- Need this because we want the robot to learn to corelate from visual features with its own joint positions.
- Joint positions:
- This is basic.
Observation.joint_position()
- This is basic.
- Final Position Coords:
- This is because the agent should learn to make decisions knowing whats the final goal positions.
task.target.get_position()
- This is because the agent should learn to make decisions knowing whats the final goal positions.
- RGB values of the image showcasing robot from multiple positions at each timestep:
todo
todo
- Number of Steps to Goal Position post Learning.
- Rewards in RL setup