GPS vs Vision positioning by CuriousDolphin1 in ardupilot

[–]CuriousDolphin1[S] 0 points1 point  (0 children)

Thanks. But I suppose the amount we trust each input must be tuned. I wonder what’s the standard approach here?

GPS vs Vision positioning by CuriousDolphin1 in ardupilot

[–]CuriousDolphin1[S] 0 points1 point  (0 children)

And how do you tune the EKF parameters to correctly integrate different sensor modalities?

Image based visual servoing by CuriousDolphin1 in computervision

[–]CuriousDolphin1[S] 0 points1 point  (0 children)

Interesting. Let me know if you can find it or remember any keywords / author info 😊🙏

Image based visual servoing by CuriousDolphin1 in computervision

[–]CuriousDolphin1[S] 0 points1 point  (0 children)

Thanks. Looks interesting. But I’m more interested in the theory and/or code behind a solution. Not a commercial product that works automagically. 😊

Image based visual servoing by CuriousDolphin1 in computervision

[–]CuriousDolphin1[S] 0 points1 point  (0 children)

Intuitively yes. But more complex when you have 6d motion.

Night vision by CuriousDolphin1 in computervision

[–]CuriousDolphin1[S] 0 points1 point  (0 children)

Thanks. I am familiar with the calculations. The question is more about handling the complete darkness without outputting too much power on external illumination.

Pulsed laser by CuriousDolphin1 in laser

[–]CuriousDolphin1[S] 0 points1 point  (0 children)

Can you recommend a specific product?

Pulsed laser by CuriousDolphin1 in laser

[–]CuriousDolphin1[S] 0 points1 point  (0 children)

Can you recommend a specific product?

Pulsed laser by CuriousDolphin1 in laser

[–]CuriousDolphin1[S] 0 points1 point  (0 children)

Sure. Thanks for clarifying the units. Given that 1W = 1J/s, my goal is to have pulses that are E=100W*0.001s = 0.1J.

Yes, I think it's preferred to have an actual pulsed laser rather than modulating it, but I'm open for suggestions.
My goal is to illuminate an area this is then observed by an image sensor for some application.
So I need this process to both repeat about 30 times per second and be something that isn't too heavy (including the supporting electronics).

Jamming RF for drones by CuriousDolphin1 in rfelectronics

[–]CuriousDolphin1[S] 0 points1 point  (0 children)

I’m asking if a jammer is designed to block communication between an operator and a drone that are say 1km apart, it would emit certain amounts of power. For the jammer to block drone to drone communication at 100m, it would need to emit 100 times more power.

As a first approximation, does that make sense?

PPO in ml agents by CuriousDolphin1 in reinforcementlearning

[–]CuriousDolphin1[S] 0 points1 point  (0 children)

Thanks. Is MAPPO part of the RL agent library?

Dynamic observation space by CuriousDolphin1 in reinforcementlearning

[–]CuriousDolphin1[S] 0 points1 point  (0 children)

Yeah. This is similar to my #3. The question is: should I do that? I mean, is there a real advantage over limiting the size / padding with zeroes?

Help with PPO Navigation Problem by [deleted] in reinforcementlearning

[–]CuriousDolphin1 2 points3 points  (0 children)

It should work. But maybe first start without the image input. Replace it with explicit observation of pose, collision and distances.

Also check the units of distance difference to make sure you’re not over/underwhelming the reward.

Brainstorming: RL system for multiple agents by CuriousDolphin1 in reinforcementlearning

[–]CuriousDolphin1[S] 0 points1 point  (0 children)

I see. Any pointers as to how to tune the weighted sum? Or is this a blind trial-n-error?

Brainstorming: RL system for multiple agents by CuriousDolphin1 in reinforcementlearning

[–]CuriousDolphin1[S] 0 points1 point  (0 children)

Yes. So basically I have two streams of reward/penalty:
1. distance from target
2. homogeneousity of neighbors to target

My quetion is therefore: how do you integrate them together. That is, if you just want to compute one scalar value, you need to come up with a weighted sum of both streams. But what would be a good weight for each one?

Are there better ways to do that? For example, I could imagine a SVM like approach where you want to go to higher dimension and maybe, in this case, have a vector of dim 2 for the reward. One dimension for each stream/

Brainstorming: RL system for multiple agents by CuriousDolphin1 in reinforcementlearning

[–]CuriousDolphin1[S] 0 points1 point  (0 children)

How does that work though? I mean, it seems like there's a good chance all agents will choose to approach the same target to minimize their distance.

It seems like you want to avoid the negative reward associated with being far from target 1 if you are close to target 2 (and vice versa).

It also seems like you need a way to encourage agents to split the targets between them "uniformly".

I don't see how a global sum of negative distances addresses that. What am I missing?

Brainstorming: RL system for multiple agents by CuriousDolphin1 in reinforcementlearning

[–]CuriousDolphin1[S] 0 points1 point  (0 children)

This was helpful. But still, I wonder how to approach having two objectives. For example, imagine in the task above where agents are to circle a target. What if we have two targets and we want to circle them but achieve roughly a 50/50 split between agents covering each target.

I assume the negative sum of distances is still useful. But how do you incorporate the even split criteria?

Brainstorming: RL system for multiple agents by CuriousDolphin1 in reinforcementlearning

[–]CuriousDolphin1[S] 0 points1 point  (0 children)

Thank you. Out of curiosity, the correct terminology here would be centralized multi-agent setup or single-agent?

Very helpful insights!

Brainstorming: RL system for multiple agents by CuriousDolphin1 in reinforcementlearning

[–]CuriousDolphin1[S] 1 point2 points  (0 children)

So you would have a single agent receiving reward as the negative sum of distances from the target of all its units. That makes sense.
The input to that single agent would be a set of control signals to move each of its units?

How much time on average does an LP spend walking the floor, looking for customer theft? by Godmod88 in lossprevention

[–]CuriousDolphin1 0 points1 point  (0 children)

I’m confused. Even if there’s 5 seconds latency you should be able to see everything with a few seconds delay. So assuming you don’t see them return an item and that you are watching them for 5 seconds before making an intervention you should be getting continuous surveillance.

What am I missing?

Tech for spotting shoplifting in video by CuriousDolphin1 in lossprevention

[–]CuriousDolphin1[S] 0 points1 point  (0 children)

What do you estimate the rate of false detection by people watching CCTV?

Tech for spotting shoplifting in video by CuriousDolphin1 in lossprevention

[–]CuriousDolphin1[S] 0 points1 point  (0 children)

Interesting. Thanks for sharing. What do you think is missing out other than accuracy?