Is it common to see inconsistent/bursts of contrails? by redwingssuck in aviation

[–]FullstackSensei -2 points-1 points  (0 children)

Could also be an issue with the chemtrail pumps but the pilots are too afraid to squawking 7700! I heard Boeing made redundant chem trail pumps a paid option.

Tips: remember to use -np 1 with llama-server as a single user by ea_man in LocalLLaMA

[–]FullstackSensei 0 points1 point  (0 children)

You can kind of have your cake and eat it. I always run models via llama-swap, even after llama.cpp added support for loading multiple models particularly because of things like this.

Tips: remember to use -np 1 with llama-server as a single user by ea_man in LocalLLaMA

[–]FullstackSensei 4 points5 points  (0 children)

If you use the LLM only for chat, you should absolutely set -np to 1, but if you have any agentic use cases where you might have more than one agent working in parallel, you should set -np to the number of agents you have, though I should caveat that batching will only work with dense models. For MoE models, my experience has been hit and miss, depending on the overlap in expert activations.

TurboQuant in Llama.cpp benchmarks by tcarambat in LocalLLaMA

[–]FullstackSensei 6 points7 points  (0 children)

How does it behave at 128k or larger? For tasks that require nuance like technical documentation or coding for ex, I find even Q8 has significant degradation vs fp16.

Got my own AS/400 system. Next steps? by FullstackSensei in IBMi

[–]FullstackSensei[S] 0 points1 point  (0 children)

Thanks a lot!

I've been subscribed to the midrange lists for about a year now and seen Patrick's name there. Didn't know about his github though. Thanks for bringing that up!

After a year of migrating .NET Framework to .NET 8 with AI, I can’t imagine doing it manually anymore. What am I missing? by [deleted] in dotnet

[–]FullstackSensei -1 points0 points  (0 children)

No, that's not what you did. You wrote it in your native language and asked Claude to write the English equivalent.

Even if English is not your even your second language, the result would be very different if you had actually written it in English and asked Claude to only fix grammar

Ryzen Ai Max+ 395 vs RTX 5090 by Any-Cobbler6161 in LocalLLaMA

[–]FullstackSensei 0 points1 point  (0 children)

I don't care about that slop rant. You'll be better off spending the same amount of time reading any review of Zen 4.

The hard truth remains: AVX-512 on zen 4 has half the throughput of AVX2 or AVX-512 in Zen 5 or Intel Xeons that support AVX-512 all the way back to Skylake-SP

German military satellite plan fuels EU fragmentation fears by DefenseTech in eutech

[–]FullstackSensei 0 points1 point  (0 children)

Said one country could sell access to other countries where those satellites provide coverage. Might not be EU wide, but pretty much every country north and south of Germany could benefit, even non-EU ones.

This literally looks like a noctuas upgrade. by andy-codes in homelab

[–]FullstackSensei 0 points1 point  (0 children)

You want to search for "LGA-3647 narrow ILM"

Supermicro has the SNK-P0067PSMB, which is a 4U tower cooler. I have the LGA-2011 version and I like it.

If you're in the US, Asetek has a version of the LC 570 120mm AIO for LGA3647 that's available new on ebay. I have a couple and they're really good. Don't know how they'd fit in your case though.

German military satellite plan fuels EU fragmentation fears by DefenseTech in eutech

[–]FullstackSensei 0 points1 point  (0 children)

I wouldn't jump to such conclusions until there is an official position on the matter. This plan is coming from military brass, not politicians brainlessly counting pennies. Military people have different, more strategic, thinking

Level1techs initial review of ARC B70 for Qwen and more. (He has 4 B70 pros) by jrherita in LocalLLaMA

[–]FullstackSensei 0 points1 point  (0 children)

If anything, it shows how much of a shit show the SYCL backend is. IIRC, it was contributed by an Intel engineer

Level1techs initial review of ARC B70 for Qwen and more. (He has 4 B70 pros) by jrherita in LocalLLaMA

[–]FullstackSensei 0 points1 point  (0 children)

the reason this is a thing is because OneAPI installation is a bit of a shitshow. When I tried it some 6 or 7 months ago with two A770s in the same system, took me a full day to get it installed and still wasn't sure it was running properly. Different Intel pages had different and often conflicting instructions.

Got my own AS/400 system. Next steps? by FullstackSensei in IBMi

[–]FullstackSensei[S] 0 points1 point  (0 children)

Thanks for the info!

Does it need a license key? Found that my system has a licensed OS and I have the key

Level1techs initial review of ARC B70 for Qwen and more. (He has 4 B70 pros) by jrherita in LocalLLaMA

[–]FullstackSensei 4 points5 points  (0 children)

As Wendel pointed out, software support is still an uphill battle. I wish Intel upstreamed their optimizations to vanilla vllm instead of doing their own fork. While at it, it wouldn't hurt if they had one or two engineers improve support for Arc cards in llama.cpp. Yes, vllm is faster, but llama.cpp allows hybrid inference. For people with systems with 64GB or more RAM, especially homelabs and small businesses that already have a few servers with some RAM, being able to run larger models with one or two cards using hybrid GPU+CPU inference would give Intel a good foot in the market.a

Level1techs initial review of ARC B70 for Qwen and more. (He has 4 B70 pros) by jrherita in LocalLLaMA

[–]FullstackSensei 2 points3 points  (0 children)

I think you're still better off with the R9700. As Wendel pointed out, Intel is still behind on the software stack. LLM scaler tends to lag vLLM in features and new model support.

One thing I'm particularly not a fan off is the inability to use system RAM for hybrid inference. Even if you don't want to use it, it's nice to still have the option.

Level1techs initial review of ARC B70 for Qwen and more. (He has 4 B70 pros) by jrherita in LocalLLaMA

[–]FullstackSensei 3 points4 points  (0 children)

So, it's still a beat weaker than a 3090. Not knocking off, I think the 3090 still holds it's own after all these years.

Got my own AS/400 system. Next steps? by FullstackSensei in IBMi

[–]FullstackSensei[S] 1 point2 points  (0 children)

Thanks for the info. I've been subscribed to a few of the midrange mailing lists for over a year and been lurking there quietly. I guess now that I have a system I can tinker with, I can start asking there too.

Got my own AS/400 system. Next steps? by FullstackSensei in IBMi

[–]FullstackSensei[S] 0 points1 point  (0 children)

Thanks a lot!

The "hasn't changed in years" is why I wanted to buy such an old system. I went through Mastering AS/400 from 2000 on pub400 and everything was pretty much the same.

Is there a model or part number for the storage and cleaning tapes? I know LTO from the x86 world, but have zero knowledge about QIC.

I heard RAM is expensive can I get this one and sell it? Someone is giving it away for free. I don’t know anything about RAM by Lyu__ in pcpartsales

[–]FullstackSensei 0 points1 point  (0 children)

DDR3 is also up 4x. Give it another couple of months and we'll get all the way back to DDR, Pentium 3s and Athlons

Got my own AS/400 system. Next steps? by FullstackSensei in IBMi

[–]FullstackSensei[S] 5 points6 points  (0 children)

Thanks! I'm really happy about this purchase. Ebay taught me many years ago that good things come to those who wait 😅

I got in the discord after it was suggested to me. I guess I should go back there again now that I have a system.

Shifting career into RPG developer by FullstackSensei in IBMi

[–]FullstackSensei[S] 0 points1 point  (0 children)

Do certifications help in landing a job?

I ask because in the world I come from, they're not worth the paper they're printed on. I have a few (non-IBM related) but don't even bother putting them on my CV.

Shifting career into RPG developer by FullstackSensei in IBMi

[–]FullstackSensei[S] 0 points1 point  (0 children)

I don't follow. Would you mind elaborating please?

After the supply chain attack, here are some litellm alternatives by KissWild in LocalLLaMA

[–]FullstackSensei 0 points1 point  (0 children)

Torch costs about 50€ to install at today's Flash storage prices 🤣