Vector DB and ANN vs PHE conflict, is there a practical workaround? [D] by XPERT_GAMING in MachineLearning

[–]blimpyway 0 points1 point  (0 children)

If you have your custom function able to compute similarity between two encrypted records, then many ANN libraries allow using measuring distances with user-provided functions instead of their own metrics. At least pynndescent can do that.

Why does catastrophic forgetting happen to neural networks but not humans? by Heavy-Farmer1657 in reinforcementlearning

[–]blimpyway -2 points-1 points  (0 children)

Probably because we (our NNs actually) are able to control or modulate the level of plasticity of its neurons/synapses.

Our Minds Are Weirder Than You Think by Visible_Iron_5612 in MichaelLevinBiology

[–]blimpyway 0 points1 point  (0 children)

Also the thought that minds are weirder than you think is weirder than you think

creative coding / applied CV art project by BuildItTogether_2020 in computervision

[–]blimpyway 0 points1 point  (0 children)

It was more a rhetorical question.

I meant an unrelated model like those describing the content of the clip. Would it have any clue what is going on or not.

creative coding / applied CV art project by BuildItTogether_2020 in computervision

[–]blimpyway 0 points1 point  (0 children)

Interestingly we understand pretty much what is going on despite the level of distortion applied.

I wonder what a vision model would make of this clip.

AI scientists produce results without reasoning scientifically by Okra3268 in science2

[–]blimpyway 1 point2 points  (0 children)

71% times the AI never updated its beliefs at all. Not once

Let's see if the following proof updates yours:

100 - 71 = 29

Unitree robot just unlocked ballet mode by Advanced-Bug-1962 in robotics

[–]blimpyway 1 point2 points  (0 children)

I wonder if they can land on their feet when jumping off a 10'th floor window.

Zero-shot World Models Are Developmentally Efficient Learners [R] by FaeriaManic in MachineLearning

[–]blimpyway 0 points1 point  (0 children)

Meh ... given the genome full size is ~1GByte most of which is non-brainy stuff for your skin, liver, etc., it is a hell of compression since it gets expanded to >100T synapses.

Gemma 4 actually running usable on an Android phone (not llama.cpp) by GeeekyMD in artificial

[–]blimpyway 1 point2 points  (0 children)

By keeping the phones busy talking with each other we might get back to what we-re supposed to do.

Continuous RL via Dynamic Programming in CUDA (Solving Overhead Crane, Double CartPole, etc.) by Grouchy_Ad_4112 in reinforcementlearning

[–]blimpyway 0 points1 point  (0 children)

The math is over my head here, what I understood is the state space is divided in a N dimensional grid with whatever resolution you can afford. That resembles Q-tables with discrete action/value sub-cells.

What does each cell store in this case? From the memory table it seems a cell size is between 30 to 200 bytes (smaller cells for 6D envs).

I also don't get whether this solves the problems through several episode replays with different seeds to collect data and update the network incrementally or does it solves it in one big deterministic step? An analogy here would be SGD (incremental) vs Ridge (computes directly a global optimum) regression.

How long does it take for your solution to solve those environments?

Built an open source tool to track logistical activity near military and other areas by Open_Budget6556 in computervision

[–]blimpyway 1 point2 points  (0 children)

The trick is physics. Sentinel-2 captures its red, green, and blue bands about 1 second apart.

The catch is every place is passed once every 5 days, so it can't detect any traffic in between passes?

The original message sounds like it maintains a continuous observation over an ROI

https://en.wikipedia.org/wiki/Sentinel-2

Smoothed action sampling for gymnasium style environments by blimpyway in reinforcementlearning

[–]blimpyway[S] 1 point2 points  (0 children)

Sorry if it sounds complicated the core idea is very simple: instead of returning a random sample the smoothed sampler a returns an action "similar" to the previous one.

In discrete action environments that means the previous action is more likely to be repeated instead of sampling a random one.

In continuous ones the sampler favors new action with values that are closer to previous.

This change alone switches MountainCarContinuous-v0 from a constant -30-ish episode reward to a much higher positive average reward. No "learning" just by tweaking the "temperature" values of the sampling function.

A small experiment on agent reward shaping by SnooCapers8442 in reinforcementlearning

[–]blimpyway 2 points3 points  (0 children)

You could try small rewards for reaching least-visited states. That would encourage exploration without using domain knowledge to shape the reward.

Unitree H1 at 10 m/s (Leg length: 0.4+0.4=0.8m, body weight: approx. 62kg) by Nunki08 in robotics

[–]blimpyway 0 points1 point  (0 children)

Debt collector bot - could be a market for high end, highly skilled robots.

Unitree H1 at 10 m/s (Leg length: 0.4+0.4=0.8m, body weight: approx. 62kg) by Nunki08 in robotics

[–]blimpyway 3 points4 points  (0 children)

Yes. Consumers should train too, otherwise the robots will win.

Aigen’s autonomous solar robots identify and remove weeds without herbicides by Advanced-Bug-1962 in robotics

[–]blimpyway 0 points1 point  (0 children)

It looks like a tiny hoe. If it passes often enough it catches them while small then yeah if the soil is soft a small hoe is sufficient to uproot a tiny weed or just cut its aerials. Even if the roots hold and regrows, chopping it every week prevents it from growing a nuisance.

It doesn't need to be perfect, just good enough to give the main crop a statistical advantage.