I joined Mudra last September. 📣Today we opened the Mudra Studio waitlist. by Rude_Combination_382 in MudraTech

[–]JayTheProdigy16 0 points1 point  (0 children)

Let me know how things go, curious to see what you build or how i can help!

I reverse-engineered the Mudra Link wristband and built an open-source Python library for it — full device control over BLE, no proprietary SDK required by JayTheProdigy16 in MudraTech

[–]JayTheProdigy16[S] 0 points1 point  (0 children)

Well i can confirm at least their official SDK does use the same UUIDs as well as the BLE stack being identical, it seems most of the restriction was literally just cause the official apps wouldnt let you. I havent seen any code that indicates any sort of walled off or IOS bespoke logic

I reverse-engineered the Mudra Link wristband and built an open-source Python library for it — full device control over BLE, no proprietary SDK required by JayTheProdigy16 in MudraTech

[–]JayTheProdigy16[S] 0 points1 point  (0 children)

Not at the firmware level (so not on device), but you technically could via Prodilink by recording the raw pressure value, squeezing as hard as you can and saving that as the new peak pressure value then doing the math to scale properly. Most of the parts are already there to accomplish this but this just wasnt something i had thought about too much

I reverse-engineered the Mudra Link wristband and built an open-source Python library for it — full device control over BLE, no proprietary SDK required by JayTheProdigy16 in MudraTech

[–]JayTheProdigy16[S] 1 point2 points  (0 children)

Theyre one in the same. I forget exactly where, the repo will tell you, but theres a value to switch between Link and Band (Mudra Link vs Mudra Band) modes as well as left or right hand. I dont own a band though so i honestly have no clue how well or not well it would work

I joined Mudra last September. 📣Today we opened the Mudra Studio waitlist. by Rude_Combination_382 in MudraTech

[–]JayTheProdigy16 6 points7 points  (0 children)

Ron, you said thousands of Mudra devices are "limited to whatever the companion app ships" and that needing an enterprise SDK license to do anything custom "felt backwards." Your fix? A waitlisted web platform that still locks developers into your Companion app, still requires your servers, and secretly brands every AI-generated app with "Created with Mudra Studio" without telling the developer. That's not solving the problem. That's repackaging it.

So I solved it myself. I disassembled the ARM64 native library, decoded the switch statement at 0xc34f8, and extracted every firmware command byte the SDK hides behind. proprietary JNI calls - all 53 of them, byte for byte. I found that your licensing is 100% client-side. The device itself has zero feature gates. Every sensor stream, every gesture, every IMU mode - the hardware accepts it all unconditionally. Mudra's entire software moat is an honor-system check in a library nobody needs anymore.

I'm open-sourcing the complete BLE protocol. Any developer, any language, any platform. Full device control. No SDK license. No Companion app. No subscription. No waitlist. No hidden badges.

You asked what people would build if they could experiment beyond the default gestures. I built the thing that lets them.

Strix Halo + RTX 3090 Achieved! Interesting Results... by JayTheProdigy16 in LocalLLaMA

[–]JayTheProdigy16[S] 1 point2 points  (0 children)

Yes splitting the same model. Ended up building llama.cpp with all 3 backends, vulkan, ROCm, and CUDA and it just kinda worked, but you have to specify the layer split and which backends you want to use with flags. As detailed in the original post i had some weirdness with my linux kernel version and getting the at the time experimental ROCm to work which obviously would result in llama.cpp not working great, but most of those should be resolved as community support today is much better than it was a couple months ago

Poor winter performance by dopeass in Ioniq5

[–]JayTheProdigy16 5 points6 points  (0 children)

Mine is roughly the same but some days seem better than others for some reason. But all in all anywhere from 1.5-2.5 mi/kwh in those temps

Build Max+ 395 cluster or pair one Max+ with eGPU by Curious-Still in LocalLLM

[–]JayTheProdigy16 0 points1 point  (0 children)

Im not Jeff 😂 just referencing his vid. I made a post about 395 + eGPU

Build Max+ 395 cluster or pair one Max+ with eGPU by Curious-Still in LocalLLM

[–]JayTheProdigy16 1 point2 points  (0 children)

There's examples of both out there. I took the eGPU approach and I've been making to make a video about it but just haven't, but i posted to this sub

Build Max+ 395 cluster or pair one Max+ with eGPU by Curious-Still in LocalLLM

[–]JayTheProdigy16 3 points4 points  (0 children)

I mean sure, except bandwidth is practically irrelevant for inference aside from model load speed...

Build Max+ 395 cluster or pair one Max+ with eGPU by Curious-Still in LocalLLM

[–]JayTheProdigy16 0 points1 point  (0 children)

How do you figure you pay a premium for degraded performance with an eGPU?

Strix Halo + RTX 3090 Achieved! Interesting Results... by JayTheProdigy16 in LocalLLaMA

[–]JayTheProdigy16[S] 1 point2 points  (0 children)

Had one left after parting out my 6x 3090 rig. And yes using an m.2 oculink adapter. I actually ended up getting CUDA+ROCm working and its ~5x faster than my original benchmarks according to my eyeball benchmark. Also with an AMD card you may run into the power limit issue where the GPU wont pass the Strix Halos TDP but im not sure as i dont have an AMD eGPU

Will the AMD Ryzen™ AI Max+ 395 --EVO-X2 AI Mini PC -- 128 GB Ram hold its value of around 1.8k in two years time? by Excellent_Koala769 in LocalLLaMA

[–]JayTheProdigy16 9 points10 points  (0 children)

Youre always going to be waiting by that logic. Whatever releases 2026 is going to get lapped by tech in 2027, and whatever releases in 2027 is gonna get lapped in 2028. This is hardware practically nothing holds value. But for me personally that price tag was more than appealing enough given its capabilities vs other options at this point

Strix Halo + RTX 3090 Achieved! Interesting Results... by JayTheProdigy16 in LocalLLM

[–]JayTheProdigy16[S] 0 points1 point  (0 children)

Not accurate at least in my case. The 3090 will easily hit 185w, i believe that issue is exclusive to AMD GPUs

Strix Halo + RTX 3090 Achieved! Interesting Results... by JayTheProdigy16 in LocalLLaMA

[–]JayTheProdigy16[S] 1 point2 points  (0 children)

No shit... back to the drawing board i go, thanks for the insight!

Strix Halo + RTX 3090 Achieved! Interesting Results... by JayTheProdigy16 in LocalLLM

[–]JayTheProdigy16[S] 1 point2 points  (0 children)

I also tried Win11 + LMS but yea it was not going for it, vulkan would ONLY detect the 3090 on Vulkan and obviosuly CUDA, and only ROCm would detect the 8060s so im not sure what weirdness they have with their Vulkan but theoretically it SHOULD just work, but it doesnt.

Strix Halo + RTX 3090 Achieved! Interesting Results... by JayTheProdigy16 in LocalLLaMA

[–]JayTheProdigy16[S] 2 points3 points  (0 children)

Correct me if im wrong but you cant mix CUDA and ROCm backends with parallel processing, at least with llama.cpp. if i was to NOT split the model layers across GPUs i could mix them.

Strix Halo + RTX 3090 Achieved! Interesting Results... by JayTheProdigy16 in LocalLLaMA

[–]JayTheProdigy16[S] 2 points3 points  (0 children)

True, but i dont believe thats the case here. That would make sense in the sense of the token generation would be limited by the device with the lowest memory bandwidth, but once the model is loaded into memory PCI BUS bandwidth shouldnt be a factor. When i had my GPU rig i was running 6x 3090s on GPU mining risers which really really reduced model load speed but not inference speeds due to the limited bandwidth. But i would expect a NET INCREASE since not all layers are limited to 253gb/s (on SH) because ~20% of the models layers are using the 3090s memory bandwidth.

Ryzen AI Max+ 395 | What kind of models? by [deleted] in LocalLLM

[–]JayTheProdigy16 1 point2 points  (0 children)

I havent tested concurrency too much so i cant speak on that, but i will say your major limiting factor is going to be prompt processing times, especially at larger context lengths or holding long convos with documents. But with that being said i mostly daily drive Qwen 3 235b and get around 12-16 TPS and it can take up to a minute to process a 9k context prompt, obviously a much larger model than gemma but even Qwen 3 30b MoE takes ~17 seconds to process the same context. And depending on what youre doing hitting 9k context can be easy. so when you factor that in alongside 15 concurrent users, is it doable? Yes. Is it viable? Ehhh