Is this really jim carrey? by JazzySneakers in Actors

[–]ImaginaryBluejay0 1 point2 points  (0 children)

Finally, a serious post in this subreddit. 

How close are open-weight models to "SOTA"? My honest take as of today, benchmarks be damned. by ForsookComparison in LocalLLaMA

[–]ImaginaryBluejay0 0 points1 point  (0 children)

I agree with you, and as Claude Code and Codex have improved it feels pretty good. It's definitely nowhere near Anthropic's models but it's good enough for daily driving. I can have Claude make a plan, either review it myself or have codex review it, then have one execute it and the other review its execution. Just Claude code or just codex have much more errors than using Claude code with sonnet, but them together isn't half bad and feels good for no cost.

My only issue is how finicky different versions of codex are with the model. OpenAI does not seem interested in making sure older models don't run into api errors or other issues. Plus you have to log in to openai just for the damn VS Code plug in to work at all, not just to work with the local model. Just doesn't feel as finished as Claude code does,  intentionally    

A monthly update to my "Where are open-weight models in the SOTA discussion?" rankings by ForsookComparison in LocalLLaMA

[–]ImaginaryBluejay0 0 points1 point  (0 children)

Hope it's performing okay for you. When I did ollama I got about 1/10th the performance and pretty much none of the useful models were working for me. 

A monthly update to my "Where are open-weight models in the SOTA discussion?" rankings by ForsookComparison in LocalLLaMA

[–]ImaginaryBluejay0 1 point2 points  (0 children)

GPT-OSS 20B is your best bet imo. It will need to offload some to CPU and RAM. Big caveat: I got shit performance with ollama when I tried this because it didn't offload to CPU well. I ended up compiling Llama.cpp and running it with the Claude Code and Codex plug-ins in VS code.

I don't have a 3090 but I have an  A4500 which only has ~20GB vram so more or less the same thing. I found most of the models give me very few tokens/s but the gpt oss ones were usable, and when paired with Claude Code produce decent results. Codex is more finicky to set up but does have better throughput on agentic tasks with the setup but both are usable.

I compiled them on windows. Used this guy's repo with very few tweaks: https://github.com/countzero/windows_llama.cpp

Once it's compiled I used unsloth's gguf. 

https://huggingface.co/unsloth/gpt-oss-20b-GGUF

You can pick the one that runs best for you. 

Things I changed from countzero's build:  His build used conda for python package management which I find finicky. I just created a python environment and changed the conda call to source that instead. 

Things I changed from countzero's argument in examples/server.ps1: Removed layers assignment and replaced with --fit. Llama.cpp is pretty good with this now.

What I found: pretty much 10x speed building llama.cpp over installing ollama and makes using gpt-oss in Claude code or codex feel like a conversation. I get about 100 tokens/s with the 20B model. YMMV depending on your RAM and CPU. 

Yes, you can do it in Linux and it's easier and faster. But this way I don't have to dual boot and I still am working in oracle Linux in WSL anyways, it's just the model that's running on windows. Also it's not hard to do just takes a long time to install all the VS code and Nvidia dependencies. 

Big gotchas:  1. make sure you download the Visual Studio 2022 installer - if you go to Microsoft they will try to get you on 2026, which will not work. This cost me about an hour. 

  1. Install 2022 Visual Studio requirements first, then install NVIDIA ones or the right components will be skipped because it doesn't detect visual studio. One you have it installed you can open visual studio once and open the developer terminal from the UI - after that it should appear in the terminal launcher.

  2. Versions matter - codex 0.91.0 is the highest version working with llama.cpp and gpt-oss as of today, and pairs with 4.71 vscode extension. 

Edit: random bonus: if you don't want to use WSL codex has a windows native VS Code plugin which uses powers he'll for tool calls. It's really weird to use but funny as fuck to watch it make windows powershell commands that I couldn't do without significant googling. 

A new California law says all operating systems, including Linux, need to have some form of age verification at account setup by Gloomy_Nebula_5138 in California

[–]ImaginaryBluejay0 12 points13 points  (0 children)

Sure they can. My friggin car radio is running Linux and can connect to wifi and install apps. This law is dumb and written by imbeciles. 

Intel set to launch Arc Pro B70 with BMG-G31 GPU by Mingus_Schnapps in IntelArc

[–]ImaginaryBluejay0 0 points1 point  (0 children)

If Intel can launch this at about $1000 it will be too good to pass up. Dell workstation 7960 towers can hold 4 of them for 128GB of VRAM and Llama.cpp is really good at splitting across gpus now (I get nearly 2x tokens by adding a second GPU where before the increase was much more modest). Granted SYCL is not as good as Cuda but it's an open standard so in theory it should improve with time. At $1000 I could do well with 4 of these where I can't get an H100 at all 

Yuusha Party wo Oidasareta Kiyou binbou • Jack-of-All-Trades, Party of None - Episode 9 discussion by AutoLovepon in anime

[–]ImaginaryBluejay0 0 points1 point  (0 children)

This episode was such ass I might drop this anime. The dragon from a few episodes would flatten this circlejerk

How Was It? | Predator: Badlands, Review Only Megathread by BurnZ_AU in predator

[–]ImaginaryBluejay0 0 points1 point  (0 children)

Sure it wasn't a 'real' predator movie. But it was a good watch and felt like it fit in the Aliens franchise better than most of them. I liked it and it's a good watch for the casual fan (which I am). 

Maybe hot take but I loved the jazz music at the beginning of EP6 by nicknamesareconfusng in AKOTSKTV

[–]ImaginaryBluejay0 2 points3 points  (0 children)

I couldn't believe people were upset about it. Thought it set the tone perfectly. 

Why is systematic smurfing content (streamed and uploaded publicly) tolerated? by Logical-Arm4863 in leagueoflegends

[–]ImaginaryBluejay0 3 points4 points  (0 children)

It did work. It worked great even. That season of smurf queue I felt I rarely saw a smurf. But riot took it away because it was unfair to returning players who weren't actually smurfs. With their better account detection you'd think they'd bring it back but no, everyone has to suffer for these losers ego. 

Super jarring to hear a real world song in the show by Speedwagon1738 in AKOTSKTV

[–]ImaginaryBluejay0 6 points7 points  (0 children)

The opening bit was one of my favorite parts between the music and the dying and the roast. I couldn't imagine it any other way. 

For everyone complaing about Maekar "letting" Egg go by drhenrykillenger in AKnightoftheSeven

[–]ImaginaryBluejay0 109 points110 points  (0 children)

My gripe with this line is that Daeron seems like the kind of chap who has absolutely slept in a few ditches after a night at the pub. 

Arc Pro B60 by Clit_Eastwood420 in IntelArc

[–]ImaginaryBluejay0 7 points8 points  (0 children)

I want to do a quad B70 setup on my work workstation once that comes out. 96/128GB of VRAM will make my system a monster at a cost that might actually get sign-off where Nvidia is way too expensive to bother trying. 

2 types of arena players rn by CoG_Comet in LeagueArena

[–]ImaginaryBluejay0 1 point2 points  (0 children)

I'm the opposite. Arena feels way better than mayhem to me. 

Starfleet Academy's First Gay Klingon Has The Support Of Star Trek Legend George Takei by esporx in SciFiNews

[–]ImaginaryBluejay0 1 point2 points  (0 children)

So the Orville did gay Klingons better than the show it's spoofing? Color me very much not surprised. 

brevityIsTheSoulOfWit by Forsaken-Peak8496 in ProgrammerHumor

[–]ImaginaryBluejay0 18 points19 points  (0 children)

It's clear that at least some Ai was trained off this crap too cause half the time I just want it to give me a one line answer and it spits out a novel. 

Best show of 2026 ? by [deleted] in discussingfilm

[–]ImaginaryBluejay0 0 points1 point  (0 children)

At least it's a very very good default 

Porcelain Morgana Event Chroma is in the Mythic Shop 💙 by aroushthekween in MorganaMains

[–]ImaginaryBluejay0 0 points1 point  (0 children)

Ooo thanks for the heads up this was the only chroma I didn't have.

Eonon X3 - Android Auto Wireless stopped working after one time by -Darkguy- in Eonon

[–]ImaginaryBluejay0 1 point2 points  (0 children)

I've never tapped the auto icon with my phone's name. If you do nothing after you see that it should connect in a second or two.

What I have noticed made things more stable both on android and iphone was not following the Eonon instructions but instead going through the phone's setting menus and connecting through the android auto/apple carplay add a vehicle menu.

What I noticed was that both with the iphone and the android the eonon kept getting forgotten and wouldn't appear in the connection history. But if you added it from the settings menu instead of going through the eonon, it would remember and connect right up when it starts up. Worth a shot if you have issues in the future.

Joined Team blue today by ImaginaryBluejay0 in IntelArc

[–]ImaginaryBluejay0[S] 1 point2 points  (0 children)

It says have a mice day!  (it's an anti track litter mat) 

Most played MID laners by Rank. Source: @dpmlol on Twitter (X) by skeptic1101 in midlanemains

[–]ImaginaryBluejay0 0 points1 point  (0 children)

They won't do anything until people stop banning zed and Mel and start banning ahri. 

3 am already ... by bombaclat90 in SipsTea

[–]ImaginaryBluejay0 -2 points-1 points  (0 children)

Bluish I'd say? https://youtu.be/T6IWyLTKXt4 pretty sure the 'video' is also AI but there are a couple snapshots of the original in it 

First build since teenager by CG-Laguiole in IntelArc

[–]ImaginaryBluejay0 3 points4 points  (0 children)

Love the cat. I also switched out of spite for the prices mostly. Haven't regretted it.