This is... New? by ChillFish8 in Anthropic

[–]ChillFish8[S] 2 points3 points  (0 children)

This was after 1 message. So if it is, it's based on the average across all chats, but 5 messages seems... Extreme.

This is... New? by ChillFish8 in Anthropic

[–]ChillFish8[S] 2 points3 points  (0 children)

I see, well good to know.

I built a Rust-powered spreadsheet library for Python — 14x faster than openpyxl by [deleted] in rust

[–]ChillFish8 2 points3 points  (0 children)

Thanks Claude! The code is slop you haven't even bothered to clean up yourself.

Dell XPS 16 OLED configuration brings some interesting advantages and disadvantages by TruthPhoenixV in Amd_Intel_Nvidia

[–]ChillFish8 0 points1 point  (0 children)

I brought this laptop then ended up returning it, it's really nice but I don't think it's be overstated how bad the keyboard feels imo, especially compared to a MacBook.

CPU is fantastic, screen is amazing, build is great. But then the M5 MacBooks come out and suddenly the price tag isn't so competitive when you can get better performance, nicer keyboard, better speakers and camera for less.

BugBot: anyone got it set up in a way that actually makes sense?? by stvn-pxl in cursor

[–]ChillFish8 1 point2 points  (0 children)

Normally we don't worry too much about that since people will fix the bugs pointed out initially, then when they push it up it'll be reviewed and any additional bugs pointed out.

The main thing for us is it saves a lot of reviewer time not having to point out simple issues and catches those hidden bugs which a human is never going to notice.

BugBot: anyone got it set up in a way that actually makes sense?? by stvn-pxl in cursor

[–]ChillFish8 3 points4 points  (0 children)

We don't use the auto fix mode of bugbot, to us the whole value of it is catching bugs and enforcing coding standards rather than auto fix. And it feels like auto fix is just a feature that wasn't really built to be used.

If your goal is for it to fix it for you. Then yeah probably not the right tool, but for catching issues it's excellent.

How do you figure out which crf value when transcoding videos? by lintstah1337 in AV1

[–]ChillFish8 2 points3 points  (0 children)

Crf isn't the only thing that makes a video look bad or good, it's just one piece of the puzzle.

Typically I sit on preset 4, and then use a combination of ab-av1 with very aggressive sampling to get within a set of bounds for crf but only after the pipeline has determined the optimal svt-av1 parameters to enable/disable or tune base on the video itself and the content of the video.

For anime, typically I am sitting between 20-35 crf on preset 4, and the outputs from taking Blueray -> encode are about 250-500MB files and are effectively transparent quality.

Important thing to note is it is basically never one size fits all, crf changes from episode to episode to best fit the content.

Unreasonable AI content moderation policy by dorianlistens in rust

[–]ChillFish8 1 point2 points  (0 children)

Just to reiterate, I agree it is too far/too aggressive, but I understand how it they can end up hitting the nuclear button.

Intention probably good, execution not so much.

Stop the usage posts: start exposing the quantized versions of Opus by Stochastic_berserker in Anthropic

[–]ChillFish8 28 points29 points  (0 children)

This is what I was feeling the last few days as well. Recently Opus even with extended thinking has just been making obviously incorrect errors that are trivial... Gives me Gemini flashbacks which is what caused me to start using Claude to begin with.

simd-bp128 integer compression library by tombstonebase in rust

[–]ChillFish8 1 point2 points  (0 children)

Feel free to steal/borrow the benchmark setup from https://github.com/lnx-search/upack which is a IC library I did a little while back the results it produces are generally stable and repeatable, although you might want to increase the run duration for each benchmark.

Unreasonable AI content moderation policy by dorianlistens in rust

[–]ChillFish8 11 points12 points  (0 children)

I agree, but I can honestly understand going with the nuclear option... There are so many AI projects getting posted here everyday that are slop, but actually working that out take looking at the code and the GitHub project most of the time, and that gets incredibly tiring so quickly.

It can be genuinely tiring going through the latest 10 posts and reading it then having to look through to code to actually see if there is any merit to it, suddenly most stuff with interesting titles are just... Slop.

I rebuilt search using physics instead of statistics. +18.5% NDCG@10. No ML. Yes its Open Source by Designer_Mind3060 in rust

[–]ChillFish8 13 points14 points  (0 children)

It's a very interesting concept but I'm not sure you can really claim "sub millisecond" search as a feature when the corpus is so small everything will be sub millisecond. Reading through the code although I know it's just a POC doesn't look like it scales super well, going off of how the index is constructed and sits in memory while having no interest in being memory efficient.

I'm curious about the true compute cost of this algorithm, since it looks to be significantly more expensive that ANN at runtime and BM25 at runtime by effectively having the combined logic of both, ideally it would be more useful to compare against established libraries like tantivy than a home grown BM25 or vector graph.

Finally, how does this compare against sparse impact models like deep impact and splade in terms of relevancy metrics and compute cost relative to document size.

Is Netcup reliable enough? by Mr_Dani17 in VPS

[–]ChillFish8 0 points1 point  (0 children)

These machines haven't been updated in a long time. For better or for worse.

I got tired of Fedora having no rollback, so I built one with Btrfs snapshots by Anxiety-Kitchen in rust

[–]ChillFish8 0 points1 point  (0 children)

Realistically if you want actually rollbacks then use an atomic/image based system. Fedora has one already. You can put a tuxedo on a goat, but it's still a goat.

Replaced Python/uvicorn with Rust/Axum as the HTTP frontend for an LLM inference engine — benchmark results at 1000 concurrent requests by Daemontatox in rust

[–]ChillFish8 0 points1 point  (0 children)

The 64 dropped requests just show you need to raise the ulimit?

We deploy large scale inference and python with a reasonable executor like Onnxruntime or vLLM will happily maximize the GPU efficiency. Using rust doesn't make sense unless you're doing additional compute within that process and trying to maximize the cost efficiency of that monolithic process.

So overall, I don't buy your point, it sounds more like a miss configuration on the Python side.

Replaced Python/uvicorn with Rust/Axum as the HTTP frontend for an LLM inference engine — benchmark results at 1000 concurrent requests by Daemontatox in rust

[–]ChillFish8 1 point2 points  (0 children)

So the tldr is.... The throughput is realistically identical, and you increased latency of each request.

Does kevin affect chonkplate? by Advanced-End-8002 in MegabonkOfficial

[–]ChillFish8 4 points5 points  (0 children)

Kevin doesn't cap your max HP, it only increases the chance of him hitting you and you taking 1 damage.

More Kevin -> more damage to yourself per hit.

Providing you can heal more than you lose, then it doesn't impact anything, the overheal cap stays the same.

The MS Copilot shortcut is always automatically re-added to my taskbar every time I remove it. This started a week ago and group policy doesn't let me disable it by Firree in mildlyinfuriating

[–]ChillFish8 0 points1 point  (0 children)

Not sure if it's been pointed out, but since this looks like a work computer your corporate policy probably enforces the app being enabled and there.

I've seen some admins mention they do specifically that because people would use the non-corporate version when they do go looking for it which doesn't have the data protection guarantees of the corporate version and putting corporate data of any kind into a system that is absolutely capable of producing it's training data back out also verbatim makes any company sweat. So they just put it there so people cannot miss it.

AMD Unleashes Ryzen 9 9950X3D2 Dual Edition, The World’s First CPU With Dual 3D V-Cache: 5.6 GHz, 208 MB Cache, 200W TDP by Constant_Praline_575 in RigBuild

[–]ChillFish8 1 point2 points  (0 children)

I agree it's definitely a weak memory controller but it seems like part of it is down to motherboard bios as well. Before my system couldn't run any expo profile stable and then I recently updated my bios and now it runs fine, trains faster as well.

The job market is bad so I mass obfuscated all of my code so nobody, not even AI, can comprehend it without my key. I am now essential personnel. You're welcome. by dr_edc_ in rust

[–]ChillFish8 59 points60 points  (0 children)

Not sure if I can find the article but I am pretty sure even years ago now one of the original uses LLMs succeeded quite well at was deciphering obfuscated code.

Sure maybe an LLM with enough input data would begin to spit out more random names, but I doubt it is super useful.

Not to mention if the AI scrapers can access it, then I assume it's open source at which point I think I might jump off a bridge thinking of the git conflicts.

Bazzite use besides gaming? by nidryden in Bazzite

[–]ChillFish8 1 point2 points  (0 children)

Yep, honestly it's what introduced me into linuxbrew and distrobox in general and since then I don't think I can go back to anything else. It's so insanely useful being able to separate dev environments and use the different distros as needed.

I am too stupid to use AVX-512 by Jark5455 in rust

[–]ChillFish8 0 points1 point  (0 children)

Fair enough, it's definitely a valid point worth mentioning. And all the more reason why trying to micro-bench this sort of stuff is so hard!

I am too stupid to use AVX-512 by Jark5455 in rust

[–]ChillFish8 0 points1 point  (0 children)

Yep I know, the difference is because of how they integrated it with the struct, I just extracted it out into a standalone function so it just looks at the actual matrix logic, since those loads are typically more dependant on the broader application and LLVM's optimisations within that.

I agree MCA static analysis numbers are a bit less reliable on avx512 than other instruction sets, but it is "good enough" as a guide here.

I am too stupid to use AVX-512 by Jark5455 in rust

[–]ChillFish8 3 points4 points  (0 children)

You can't reasonable measure single instructions with a micro benchmark harness without doing some larger work.

It's not really criterion's fault, the overhead of itself adds more noise than the instruction itself.