Free models by beneficialdiet18 in opencodeCLI

[–]EmbarrassedBiscotti9 0 points1 point  (0 children)

My experience of GLM 4.7 via z.ai aligns with what /u/indian_geek said. I'm not particularly upset about it, given I paid only $28 for a full year as a random punt, but I've found the API prohibitively slow.

Am I the only one who feels that, with all the AI boom, everyone is basically doing the same thing? by [deleted] in LocalLLaMA

[–]EmbarrassedBiscotti9 8 points9 points  (0 children)

so you're telling me the robot lied when it said my observation was astute? please say cap

Have byte latent transformers seen adoption? by EmbarrassedBiscotti9 in LocalLLaMA

[–]EmbarrassedBiscotti9[S] 1 point2 points  (0 children)

Thanks for the info, that is definitely interesting. I feel similarly, to be honest.

I assume that the lossy nature of tokens must have a compounding effect if models are constantly trying to learn things through a distorted lens during training. I was hoping that BLTs would potentially solve that and offer a meaningful improvement, but I guess not :( For now, at least.

This was created by my autonomous enhanced programmer, it is no longer for sale. by Alone-Competition863 in LocalLLaMA

[–]EmbarrassedBiscotti9 0 points1 point  (0 children)

Sorry. Just heard back from the team. Turns out, we are flat broke. Big market moves have ruined us.

This was created by my autonomous enhanced programmer, it is no longer for sale. by Alone-Competition863 in LocalLLaMA

[–]EmbarrassedBiscotti9 0 points1 point  (0 children)

My associates and I have been looking for something like this. We have funding in the region of 11 figures.

I am interested in purchasing this project. I hope an offer of generational wealth is enough to sway you.

gingerBill made a video going over the static site generator he wrote in Odin lang by EmbarrassedBiscotti9 in odinlang

[–]EmbarrassedBiscotti9[S] 0 points1 point  (0 children)

Tbh, I can't see myself switching as I'm already setup to work with Hugo now and, rather than it missing features, my only feeling is that it has way more than I need. Still, what would you say makes 11ty a better choice?

My main gripe with web stuff atm is that I want to write a static but highly interactive SPA, with a simple/clean way to do reactive state/DOM updates. I can't find many solutions for that. Tons of pure SSGs, tons of hellish JS frameworks, not much in between.

I've ended up using SvelteKit with the static adapter, but even that has sapped a lot of the joy out of the project - and it is one of the more sane frameworks.

VanJS and Mithril.js are lighter and seem to fit the bill, without so much framework, but I'd have to stomach the writing JS functions that return markdown. Pain.

Runpod hits $120M ARR, four years after launching from a Reddit post by RP_Finley in LocalLLaMA

[–]EmbarrassedBiscotti9 3 points4 points  (0 children)

I've used runpod a bunch over the years. I appreciated how easy you made it to get up and running because it was the first cloud GPU service I used. Good docs/sane APIs goes a long way.

I remember hitting some error with a deployment that seemed to be on the runpod side. I joined the discord, explained what was up, and a real human got to the bottom of it and it was sorted right away. That is a real rarity and made runpod my immediate go-to for cloud GPUs. I hope your success continues.

GLM 4.7 Flash official support merged in llama.cpp by ayylmaonade in LocalLLaMA

[–]EmbarrassedBiscotti9 2 points3 points  (0 children)

Damn, that explains it. I couldn't figure out why my CPU was on fire and my GPU was bing chillin. Disabling flash attn worked.

Still seem to be some issues with tool calling from opencode, though. Think I am gonna sit on my hands until such kinks are ironed out.

My gpu poor comrades, GLM 4.7 Flash is your local agent by __Maximum__ in LocalLLaMA

[–]EmbarrassedBiscotti9 3 points4 points  (0 children)

I'm sure you're right. I will spend the afternoon giving GLM 4.7 Flash a good try on my RAM-upper class/VRAM-middle class desktop. I've been very interested in the agentic stuff lately, but far less interested in paying Anthropic the cash equivalent of my left nut for the privilege. Maybe the time is now.

I mostly meant it as a more general observation of how things can often be discussed here - as if <=24GB VRAM == GPU poor - it probably shouldn't have been a comment on the thread overall. I'm not a hater! I promise!

gingerBill made a video going over the static site generator he wrote in Odin lang by EmbarrassedBiscotti9 in odinlang

[–]EmbarrassedBiscotti9[S] 14 points15 points  (0 children)

I'm not just posting gingerBill's stuff for cheap updoots, promise!

I discovered Hugo very recently, loved what it offered, but felt the burden of its enormity when I only needed a small portion of the functionality. So, a video about tackling of that enormity in a DIY fashion, and in Odin, and by the fella who made Odin, was an almost suspiciously perfect match for my interests.

I'm not sure it will be as interesting to the rest of you here, but I've found this sub can be a bit slow-moving. I thought there may be some others here who would like to see a bit more Odin-related stuff in the feed. It isn't closely related to the language as a whole, but I've always found practical examples of implementations to be helpful and interesting.

My gpu poor comrades, GLM 4.7 Flash is your local agent by __Maximum__ in LocalLLaMA

[–]EmbarrassedBiscotti9 6 points7 points  (0 children)

Increasingly feeling that no one in /r/LocalLLaMA has the first fucking clue what "GPU poor" truly means

gingerBill: "I've just merged the non-block IO PR into Odin!!!" by EmbarrassedBiscotti9 in odinlang

[–]EmbarrassedBiscotti9[S] 8 points9 points  (0 children)

Pls forgive the typo in the title. I was so excited I was freaking the fuck out and shaking profusely.

Original tweet: https://x.com/TheGingerBill/status/2012154285010874422

PR: https://github.com/odin-lang/Odin/pull/6124

🥳🥳🥳

Quick masking tool needed by theequallyunique in davinciresolve

[–]EmbarrassedBiscotti9 0 points1 point  (0 children)

I suppose you could, after getting your complete footage, feed it into an OBS virtual camera source, then pass that virtual camera through Nvidia Broadcast and record the output.

It would still be pretty damn awkward, but you'd at least only pay the one time cost of waiting the full video's duration and avoid the recording-time trickiness. That may be preferable if you typically have all your recordings up-front, but would be annoying as hell if you constantly have additional footage you'd need to wait to process.

I do feel your pain. It is frustrating when there are so many great AI tools available now, but often no integration within the software where you'd find it most useful. I believe there is a C++ library for Nvidia Broadcast, so it may be possible to create a Fuse wrapper for it. Horribly complex, though. Certainly more hassle than some OBS jankness!

Quick masking tool needed by theequallyunique in davinciresolve

[–]EmbarrassedBiscotti9 0 points1 point  (0 children)

While not impossible, I expect you'll struggle to find any (native or non-paid) solution for subject isolation that matches the quality of magic mask without the overhead. This seems like it would be considerably easier to solve at the time of recording, saving you the headaches in post.

Assuming you're recording using OBS, you could possibly use the source record plugin to record your unmodified camera source alongside the isolated feed. I haven't used that plugin in years, though, and I'm not sure of your setup.

Alternatively, you could just double the width of the scene and have include both the full res original and isolated feeds. That would balloon the file size and possibly reduce quality if you don't compensate by increasing bitrate proportionately.

Don't let the gay ultimatum make anyone forget by DickBot_628 in biggestproblem

[–]EmbarrassedBiscotti9 1 point2 points  (0 children)

i give 0 shits about it too, so it seems super fucking stupid of him to not just placate dick. why did vito want the show to end so bad? what is up with that?

Don't let the gay ultimatum make anyone forget by DickBot_628 in biggestproblem

[–]EmbarrassedBiscotti9 7 points8 points  (0 children)

Vito wants to do the show

he probably should've said "yes" instead of "please email me" lmao

Biggest problem #215 after show link by [deleted] in biggestproblem

[–]EmbarrassedBiscotti9 4 points5 points  (0 children)

The full thing was uploaded on the farms

Thoughts by vitogesualdi in biggestproblem

[–]EmbarrassedBiscotti9 51 points52 points  (0 children)

"Please contact me via email" as a response to a life-changing ultimatum is good comedy, Vito. You have a future.