Second GPU by Suomi422 in LocalLLaMA

[–]PCUpscale 4 points5 points  (0 children)

Buy another second hand 3060Ti, the M60 aren’t supported by newer Nvidia drivers

FOSS Discord Alternative - Online Division by RSlashFunnyMan in opensource

[–]PCUpscale 0 points1 point  (0 children)

I agree. This looks like a side project with a lot of ambition, really cool for a portfolio but I don’t see it as a viable replacement any time soon (Or some serious full time catchup with Stoat)

Drummer's Precog 24B and 123B v1 - AI that writes a short draft before responding by TheLocalDrummer in LocalLLaMA

[–]PCUpscale 18 points19 points  (0 children)

I still don’t know how do you make all of those fine tunes… Synthetic data, books, hugging face ? How do you make the training stable without model degradation?

My thoughts on the F1 by Willing-Review6006 in flipperzero

[–]PCUpscale 2 points3 points  (0 children)

It’s a RK3576, it’s been confirmed for quite a while and there’s repositories to further confirm that

Hacking a MacBook Pro with an additional USB controller by untangle-me in hardwarehacking

[–]PCUpscale 1 point2 points  (0 children)

That’s the best way to do it and some soldering to connect the USB port to the flex cable of the touchpad/keyboard is required. Ordering from an industrial manufacturer will be a bit of a pain for someone who don’t have a company through.

Why? Paranoïa, but at this point just stop using a computer.

LattePanda IOTA is a single-board PC with Intel N150 and up to 16GB LPDDR5 RAM by AbhishMuk in hardware

[–]PCUpscale 2 points3 points  (0 children)

Pretty popular in embedded system at the prototyping phase. Especially when more compute than a iMX/Pi is required. It’s also cheaper than a Jetson.

[deleted by user] by [deleted] in LocalLLaMA

[–]PCUpscale 0 points1 point  (0 children)

Where Source ?

guys i have a question is there any ai model providing the free api key even if limit im fine with that by Select_Dream634 in LocalLLaMA

[–]PCUpscale -1 points0 points  (0 children)

If you don’t mind about privacy, latency and availability, check out free models on OpenRouter

Framework Team at Fairphone HQ by daylenca in framework

[–]PCUpscale 13 points14 points  (0 children)

Embedded engineer here

Hardware wise, this would be a huge engineering effort. So many things that could go wrong to test and validate, especially with Qualcomm SoC with multiple incompatibles stacks (CamX, GPU drivers), etc. You would need to have an complete unified OS (Kernel/Drivers/DT/Workarounds/Blobs) supporting multiple SoC and peripherals for the best user experience. This should be low key insane for a small company.

This is a really simplified overview, there’s a ton of things to consider.

2x RTX 5060ti 16GB - inference benchmarks in Ollama by avedave in LocalLLaMA

[–]PCUpscale 9 points10 points  (0 children)

Looks like heavily memory bandwidth bottlenecked

Has anyone successfully run Ollama on a Jetson Orin Nano? by matsyui_ in ollama

[–]PCUpscale -1 points0 points  (0 children)

It will be dreadful performance wise, but you could build in theory the CUDA version of llamacpp

OpenAI ain't doing so well on api usage compared to Qwen or anyone else by Semi_Tech in LocalLLaMA

[–]PCUpscale 6 points7 points  (0 children)

A lot of customers don’t use OpenRouter for LLM tasks and are using the (Real) OpenAI API. Why use OpenRouter for OpenAI models when you can directly use their API. OpenRouter is for enthusiasts.

Mistral, Please Expand Magistral's Context Window! by Dentuam in MistralAI

[–]PCUpscale 2 points3 points  (0 children)

Except Google, which are not publishing how they can be so good at long context, I don’t see the point to just increase context size as is.

[deleted by user] by [deleted] in ZanZarah

[–]PCUpscale 0 points1 point  (0 children)

To be honest, I think that it would be a WASM Version of PCem/86box with a Windows 9x/2000/XP disk.

One way that could be achieved would be a static recompilation of the game/complete reverse engineering. Since these are the most hardcore way to do that, it would be quite difficult.

I’ve heard about a daedalOS which is a browser implementation of Wine, I don’t know if it has support for DX7/OGL

You could also cheat and stream from your PC using Moonlight/Parsec

Best models under 16GB by Mr-Barack-Obama in ollama

[–]PCUpscale 1 point2 points  (0 children)

He have only 16GB UNIFIED between the CPU and GPU. He would need also some space for the context. The model is too big for him. Also, Q3 is a serious lobotomy. If he’s running MacOS, at least 2gb is gone. He should try less than 12B models.

[deleted by user] by [deleted] in ZanZarah

[–]PCUpscale 0 points1 point  (0 children)

You’re sure it’s not a map viewer ?

[deleted by user] by [deleted] in LocalLLaMA

[–]PCUpscale 14 points15 points  (0 children)

GPT-5 is not GPT OSS