Comments (18)
So just to elaborate:
I'm assuming you are doing some RL? if that's the case, then when in this action mode:
- Your random exploration needs to be a little smarter than just random (i.e. accounting for the fact that the robot has a certain configuration space)
- Your agent also needs to be aware of this (i.e. by penalising for suggesting actions outside of this configuration space).
Hope that helps :)
from rlbench.
Hi,
Is this regarding DELTA_EE_POSE_PLAN
or DELTA_EE_POSE
? Given that you are saying: "to find a path", I'll assume you are using DELTA_EE_POSE_PLAN
. This action mode uses path planing, so yes, as with any path planning its going to take longer than simply using the DELTA_EE_POSE
or DELTA_EE_VELOCITY
, etc.
Few questions to get down to the problem:
- How are you generating these poses?
- Are you using both translation and rotation? If yes, then are your quaternion actions valid?
from rlbench.
Iām sampling x,y,z each from -1 to 1, and dividing by 100. I then use this as the delta to update.
Iām only changing x,y,z - quaternion i feed in the same from the original pose.
from rlbench.
Hi,
quaternion i feed in the same from the original pos
This is your problem. Remember, you are using the DELTA
action mode, rather than the ABS
action mode. So to get a zero rotation you need to pass in a identity quaternion. So what is probably currently happening is that you are giving a rotation that is invalid.
from rlbench.
No, I edited in the delta_ee_pose code to set the resulting action to just use the same quaternion without modifyinh
from rlbench.
If you've modified the backend then I need to see your changes. Please post the changed lines here
from rlbench.
arm_action ~ sampled from -1 to 1 each
a_x, a_y, a_z = arm_action/100
x, y, z, qx, qy, qz, qw = self._robot.arm.get_tip().get_pose()
new_pose = [a_x + x, a_y + y, a_z + z] + [qx,qy,qz,qw]
self._path_observations = []
self._path_observations = self._path_action(list(new_pose))
from rlbench.
Thanks. And how many steps does it run for before you get a (what I assume is) InvalidActionError?
from rlbench.
So If i step with [1,1,1,1(gripper)] which is [0.01...] after the division, then I get
59, in get_nonlinear_path
raise ConfigurationPathError('Could not create path.')
pyrep.errors.ConfigurationPathError: Could not create path.
on the first step.
If i step with 0.1 which is [0.001...] after the division, the error occurs on 8th step.
from rlbench.
BTW, as an aside, I believe:
self._path_observations = []
self._path_observations = self._path_action(list(new_pose))
can just be written as
self._path_observations = self._path_action(new_pose)
from rlbench.
Also, did some more digging and the 2 seconds only occurs when it can't find a path, when it does find one it's about 0.2 seconds.
from rlbench.
So If i step with [1,1,1,1(gripper)] which is [0.01...] after the division, then I get
59, in get_nonlinear_path raise ConfigurationPathError('Could not create path.') pyrep.errors.ConfigurationPathError: Could not create path.
on the first step.
If i step with 0.1 which is [0.001...] after the division, the error occurs on 8th step.
This makes sense, right? If you look at the starting configuration of the arm, the end effector cant really increase its z axis anymore without also altering the rotation, and so it will not be able to find a valid configuration. If you were to negate the z axis (which would send the end effector down), then you would find that it would run for longer before getting the error.
from rlbench.
Also, did some more digging and the 2 seconds only occurs when it can't find a path, when it does find one it's about 0.2 seconds.
This is because it keeps trying to find a path/configuration until some max_attempts limit.
from rlbench.
Ah, thanks Stephen, makes sense.
from rlbench.
Yeah, I'm trying to set up an env to match Mujoco's Fetch Reach task to baseline some earlier experiments on RLBench. So, most likely rather than configuring the agent for now I'm trying to adjust the environment to be similar. Thanks again for the help! It'd be great if you could take a look at the stepping speeds in #53 , since without faster speeds we won't be able to use this environment :(
from rlbench.
No probs š
from rlbench.
I'm not sure if this was addressed in other areas, but our project group ran into this issue and got training to run smoothly by manually forcing change in EE position to have a maximum magnitude. See here for more details.
These errors can still happen though, especially when the model has not converged. We handled two errors separately. ConfigurationPathError implies the desired action (delta EE pos) is too large. InvalidActionError implies the desired actions brings the EE outside its configuration space (ex: literally reaching beyond the arm's max reach). Please see here for how we handled this.
Hope this helps anyone facing these issues.
from rlbench.
@Alvinosaur Your team's solution seems reasonable!
Did your team have investigated the IK solvers of other RL platform like Pybullet, Robosuite, ... ?
I think addressing these IK errors occurred from explorative Cartesian space 6 DoF actions is a big challenge in robot learning domain...
from rlbench.
Related Issues (20)
- Does RLBench have instance segmentation masks?
- tools/dataset_generator.py HOT 2
- [Feature Request] Can the gym environment provide depth images and point clouds?
- Please add a detailed description of the data.
- Saving ttm issue
- utils.get_stored_demos and dataset_generator.py don't agree on mask image format, results in corrupted handles for stored demos
- How to edit obj part textures
- The robotic arm vibrates continuously when executing the example in example
- Loading demos at different resolutions causes incorrect data
- Missing actions for some timestep when using dataset_generator
- Action Execution Failed
- Why are depth images striped?
- Open box with Domain randomization at episode >=2 failed
- Forward kinematics for robot arm HOT 1
- Real time plotting issue
- File invalid Error 404
- after installing RLBench, still no module named rlbench found HOT 2
- inhomogeneous shape error when running the immitation learning example
- Can you provide obs.delta_gripper_pose in the end effector frame
- cannot reshape array of size 16 into shape (3,4)
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
š Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. ššš
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ā¤ļø Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from rlbench.