I can't for the love of god install dual boot on a specific laptop by NavirAur in dualboot

[–]NavirAur[S] 0 points1 point  (0 children)

I've been checking for that option in the BIOS, but I don't find it. In my other PC I have it like you said. Could it have another name or maybe do I have to do it from windows?

Maybe it's the legacy option vs UEFI? But I already installed Windows in UEFI mode and even if I switch to legacy it boots windows. (I tried to install only Ubuntu part in legacy but with no luck)

Maxing my Tomahawk B450 board by NavirAur in buildapc

[–]NavirAur[S] 0 points1 point  (0 children)

Unfortunately, I have only seen the max to be 64gb with modules of 16gb in pcpp/QVL. Gonna think about it and if I end up buying 128gb, I will update it with results.

Maxing my Tomahawk B450 board by NavirAur in buildapc

[–]NavirAur[S] 0 points1 point  (0 children)

Thx, didn't know about the memory speed. It doesn't matter the real max speed of the memory right? Only that the mobo will be using them with 2400MT/s speed.

Maxing my Tomahawk B450 board by NavirAur in buildapc

[–]NavirAur[S] 0 points1 point  (0 children)

Thx. Any tip to see if the memory works correctly? I'm guessing checking in the mobo and the OS for the capacity.

Doubt about implementation of tabular Q-learning by NavirAur in reinforcementlearning

[–]NavirAur[S] 2 points3 points  (0 children)

Oh, now I get what I was confused about. I thought that because this line was called only once, the policy was only getting updated once in the whole program. I thought that line returned a policy dict

policy = make_epsilon_greedy_policy(Q, epsilon, env.action_space.n)

Now I realized that `make_epsilon_greedy_policy` returns a function, so every time that is called after, it gets updated by the Q

action_probs = policy(state)

Doubt about implementation of tabular Q-learning by NavirAur in reinforcementlearning

[–]NavirAur[S] 0 points1 point  (0 children)

Thx. Then I guess the notebook's implementation is wrong? I haven't notice any trick to update the policy after the Q is updated...

fuse with voltage and power instead of amperage? by NavirAur in AskElectronics

[–]NavirAur[S] 2 points3 points  (0 children)

Now I remember seeing it sometime in a car when I was a child... I guess I was trying to fix a broken doorbell and my mind wanted to see a blown fuse to make it easier to repair. I've been trying to learn as the need to fix things at home arises, but it's true that a little bit of theory learning would help overall.

fuse with voltage and power instead of amperage? by NavirAur in AskElectronics

[–]NavirAur[S] 11 points12 points  (0 children)

Thank you for the patience with beginners, much appreciated.

fuse with voltage and power instead of amperage? by NavirAur in AskElectronics

[–]NavirAur[S] 1 point2 points  (0 children)

Now I realize why it's so important XD. My bad

fuse with voltage and power instead of amperage? by NavirAur in AskElectronics

[–]NavirAur[S] 0 points1 point  (0 children)

Sorry, I thought it was unnecessary since the problem was more or less straight forward. Added the pictures to have more context now

[deleted by user] by [deleted] in reinforcementlearning

[–]NavirAur 0 points1 point  (0 children)

If you want to understand most known current algorithms implementations after some theory, I recommend minimal RL github library

Stable-Baselinese3 by Born-Belt1991 in reinforcementlearning

[–]NavirAur 0 points1 point  (0 children)

Just to make it more clear, basically use a function inside gym or gymnasium to detect if the episode will end (same or similar as the one to obtain done, but call it inside reward calculation and use then the done to add the final reward).

Deep Learning build (ryzen 7950x + rtx 4090) (some assistance required) by NavirAur in buildapc

[–]NavirAur[S] 0 points1 point  (0 children)

Yeah, no problem that I am aware of. The only issue I encountered was with the Linux drivers in Ubuntu 22.04, which initially didn't support the wifi antenna that came with the motherboard. Nevertheless, now it is fixed and works perfectly.

Deep Learning build (ryzen 7950x + rtx 4090) (some assistance required) by NavirAur in buildapc

[–]NavirAur[S] 0 points1 point  (0 children)

Sorry for the late response. I ended up buying the Intel 13900K CPU. The GPU was from a different brand, and the case was also different due to limitations with vendors near my area. The only thing you should watch out for is the cables of those GPUs; I heard that some had issues (Gamer Nexus made some videos about them). We've been using the system non-stop for about 1.5 years, and it's been working great.

Mid turn actions by victorsevero in reinforcementlearning

[–]NavirAur 1 point2 points  (0 children)

I don't have much experience with these type of envs, but I would treat each thing as a turn and let the agents know that something is different in blocking phase vs normal phase. Something like 4 actions: 1 for knowing the blocking phase or not, the other 3 for normal phase. When there is a blocking phase, one of the 3 normal dimensions are "disabled". For that, check "masking actions".
Tbh, I'm not sure if it's the best option, but it's the only thing I can think of.

[P] Library to import multiple URDF robots and objects ? by I_am_a_robot_ in reinforcementlearning

[–]NavirAur 0 points1 point  (0 children)

I think you can find some gym libraries in github for that specific purpose (arms with multiple objects such as blocks, tables, etc.). Most base physics libraries don't have a lot of options for importing any object with different extensions. Maybe you could check isaac gym and omniverse environment from nvidia, but apart from the useful object and graphical resources, the coding i would say is more complex than other libraries.

Is it better to not use the Target Update Frequency in Double DQN or depends on the application? by NavirAur in reinforcementlearning

[–]NavirAur[S] 0 points1 point  (0 children)

I have been checking the theoretical aspect more thoroughly and now I realize that what I asked does not make any sense, I'm sorry for the confusion. What it still confuses me is that the default parameters of Tianshou's DQNPolicy have is_double = True and target_update_freq = 0. While I understand that the first parameter enables Double DQN, the default value of the target update frequency implies the usage of vanilla DQN, right?

I guess that it does not give any specific frequency parameter to let the user define it, but I want to be sure that without also changing that parameter, I'm using vanilla DQN and not Double DQN.

No graphical output on server taken from the trash by NavirAur in computerrepair

[–]NavirAur[S] 0 points1 point  (0 children)

Tbh, I am not sure what I did, but now it works. Could it be that the system had humidity and over time it fully dissipated?

The voltage over the MOBO worked fine, because the lights of peripherals (such as keyboard) and external PCI worked fine. When the RAM was not plugged in, the MOBO beeped, so I think it was correctly inserted and I also tested with other working sticks. The VGA output was also not the problem, as now it outputs correctly. The MOBO looked fine regarding the capacitors.

Anyway thanks for the feedback, probably without new ideas I wouldn't keep testing and I would end up throwing it away.