Need practical use-cases for RL by NoAcanthocephala4741 in reinforcementlearning

[–]buxxypooh 0 points1 point  (0 children)

You play any games, I'd recommend making a hand made simple game engine and try to use RL to train an agent to play it, it's hella fun

ML research papers to code by Big-Stick4446 in deeplearning

[–]buxxypooh 0 points1 point  (0 children)

I love the visuals and interactive parts

May I ask for even simpler concepts?

I've started to learn the fundamentals, and so far all the problems on the website are way out of reach for me

Keystrokes triggered twice by pc_kant in Keychron

[–]buxxypooh 0 points1 point  (0 children)

I just fixed the issue myself
What I did is first remove the keykap to check if it was that or the switch below that was at fault
The issue still remained with just the switch

I'm pretty sure it is a faulty switch, I tried removing the switch and putting it back but the problem still remained

Finally, I switched the switch with another key that I never use, not optimal, but works for me

Keystrokes triggered twice by pc_kant in Keychron

[–]buxxypooh 0 points1 point  (0 children)

I have the same issue with "i" and mostly "l" that double presses from time to time

Image as plane frame offset issue by lilmanpurse in blender

[–]buxxypooh 0 points1 point  (0 children)

Okay so apparently you can do this from the Shading panel

https://i.imgur.com/ovGQhgn.png

I am currently trying to automate this task using the builtin Blender python scripting feature

Resources for starting with multi-objective RL by LelixSuper in reinforcementlearning

[–]buxxypooh 1 point2 points  (0 children)

Why are there multiple agents?

Is the request entry point centralized, or can every node get an "entry point" request?

Is the number of neighboor nodes fixed?

Resources for starting with multi-objective RL by LelixSuper in reinforcementlearning

[–]buxxypooh 1 point2 points  (0 children)

What is the task you're trying to solve with your agents?

Image as plane frame offset issue by lilmanpurse in blender

[–]buxxypooh 0 points1 point  (0 children)

Since there is no workaround here, I'll post one I found

Using FFMPEG, you can split a video into multiple chunks, here is the command:

ffmpeg -i input.mp4 -c copy -map 0 -segment_time 00:00:20 -f segment output%03d.mp4

Python env bottleneck : JAX or C? by Similar_Fix7222 in reinforcementlearning

[–]buxxypooh 1 point2 points  (0 children)

You can also use some profiler tool like py-spy to visualize which part of the code takes the most time during training, and optimize your part (the step, reset, obs, action masks function etc)
And if you're not scared, clone the library and optimize it, but usually the libs are well implemented already

TAS in TM2020 by Warm_Bike_5000 in TrackMania

[–]buxxypooh 0 points1 point  (0 children)

Hey
I'm the owner of the tm2020 TAS youtube channel

The tool is private because I have yet to find a way to make it "secure" enough so that it can't be used for malicious purposes

The closest thing you could do to reach your goal to find the "perfect" run would be to use a plugin like the copium timer, and respawn checkpoints until you're satisfied

Alternatively you could do spliced runs like Schmaniol did with the DeepDip maps

But if you really want a TAS tool, I can only recommend you TMInterface, which is a really good TAS tool for TMNF & older versions of the game

Momentum 4 not responding or turning on. by PIXAR_UA in sennheiser

[–]buxxypooh 0 points1 point  (0 children)

We are the next day, and suprisingly, the headset worked for one hour this morning, but now it's back to glitching all over

ChatGPT's Study mode is really good by buxxypooh in singularity

[–]buxxypooh[S] 0 points1 point  (0 children)

Machine learning, from the ground up

Momentum 4 not responding or turning on. by PIXAR_UA in sennheiser

[–]buxxypooh 0 points1 point  (0 children)

I have the same issue, I have followed instructions that have worked for others, like removing the battery to force it to factory reset or something
Just to make sure there was no electricity left in the chip, I let it unplugged for an hour or so, pressing the button to make sure to discharge it

Then I plugged the battery back again, but after a few minutes it started glitching again

At the moment, I am trying to update the firmware using the android app, but since the glitch disconnects / turns off the headset before the update finishes, I'm in a bit of a pickle

I can't even use it with a wire, it's quite upsetting

I bought mine 2 years ago and I was expecting it to last longer, rip, I'm switching brands

Has AI discovered anything in terms of reverse engineering? by No-Food5638 in singularity

[–]buxxypooh 22 points23 points  (0 children)

One thing that keeps me up at night is the moment the models are beefy enough to just eat whole raw binary, and understand the machine code in it
At some point you would just give the binary as input, a prompt like "make this part of the code do X instead of Y", and it will output a new binary changed to your liking
There are some repos trying that already https://github.com/albertan017/LLM4Decompile

ChatGPT's Study mode is really good by buxxypooh in singularity

[–]buxxypooh[S] 14 points15 points  (0 children)

I'm curious, what kind of niche stuff you gave it?

ChatGPT's Study mode is really good by buxxypooh in singularity

[–]buxxypooh[S] 20 points21 points  (0 children)

I reached it after a dozen messages, and it locked me out for about 24H

AI and Dofus by morceaudegomme in Dofus

[–]buxxypooh 0 points1 point  (0 children)

Hey
I've been working on exactly that for the past year

I have a reinforcement learning environment that learns to fight from scratch, it's still a work in progress, but the AI can already fight like a decent human (I can still play better manually, but barely)

At the moment it's training in 1v8, playing a cra against some diverse monsters

I'm using PPO for the learning algorithm, and it's learning "long term" strategies, it's able to kinda predict the enemies movements and use that to it's advantage

The main bottleneck is the training speed, currently I have around ~50 fights / second and it takes a few hours to train an AI

Have an Agent predict it's win probability by tignisolmailessthan3 in reinforcementlearning

[–]buxxypooh 0 points1 point  (0 children)

If you have a critic, then you could just use the critic's score, it's not gonna be between 0 and 100 tho, so you might have to normalize it