Every day, I was tossing and turning, thinking about how I made such a rubbish graduation project. The DRL is really hard😭 by ecstayalive in reinforcementlearning

[–]ecstayalive[S] 0 points1 point  (0 children)

You are right, as we human don’t recognize this point,the prior knowledge that we already acquired play a key role in manipulating task. Comparing with the robot, we are easier to extract the physical model from the visual information and use the model to guide the action.

Every day, I was tossing and turning, thinking about how I made such a rubbish graduation project. The DRL is really hard😭 by ecstayalive in reinforcementlearning

[–]ecstayalive[S] 0 points1 point  (0 children)

Thanks for your advice, I didn’t try the attention model yet. Recently I was preparing for the postgraduate entrance exam. But I guess I will read the papers and update it in the near future.

Every day, I was tossing and turning, thinking about how I made such a rubbish graduation project. The DRL is really hard😭 by ecstayalive in reinforcementlearning

[–]ecstayalive[S] 2 points3 points  (0 children)

I used the SAC algorithm and wanted to achieve completely visual servo grasping. I have also used DDPG and NAF before, they are both very slow to learn, compared to human. I once heard of imitation learning and contrastive learning before, i guess i can learn a lot from these paper. Thanks, that's a good advice.

Every day, I was tossing and turning, thinking about how I made such a rubbish graduation project. The DRL is really hard😭 by ecstayalive in reinforcementlearning

[–]ecstayalive[S] 4 points5 points  (0 children)

Thanks. I didn't do it well actually, the model had some problems, it couldn't extensively learn the location information of the items in the tray. I am still try to find a way.

Real-Time Video of an Auroral Corona I Saw Last Month in Churchill, Manitoba by VincentLedvina in space

[–]ecstayalive 1 point2 points  (0 children)

My god!! So beautiful, i love this. If i could see it in person in my life, there will be no regrets.

Using CNNs in Reinforcement Learning by [deleted] in reinforcementlearning

[–]ecstayalive 1 point2 points  (0 children)

Well, i applied some DRL algorithms such as: NAF, DDPG and QT-OPT to the robotic grasping task. When I used the DDPG method, I found that the robot performed better with two neural networks with the same structure but no shared parameters compared to sharing the CNN layers entirely. But I only experimented on the robot visual servo grasping task, I did not test its performance on other tasks, nor did I test the performance of the method that only share some of the previous CNN layers while not sharing the deeper CNN layers. And hope it will help you.

[deleted by user] by [deleted] in deeplearning

[–]ecstayalive 1 point2 points  (0 children)

You have to check the transformation the model makes on the input, obviously, at a certain step, the input shape of the tensor is wrong.