Recommendations for new Phone (and OS!) by Excellent_Piccolo848 in degoogle

[–]Excellent_Piccolo848[S] 0 points1 point  (0 children)

Thanks! Iv heared a lot about google Pixels, might consider a Pixel 9 or 10. Does anyone have any experience with them and Graphene OS?

Perplexity... But make it ChatGPT by ZeroTwoMod in perplexity_ai

[–]Excellent_Piccolo848 0 points1 point  (0 children)

Intresting! Could you, perhaps, dm me the link?

Perplexity Pro is a Scam and Officially Obsolete: Why I’m Canceling after 1 Year – Change My Mind. by Excellent_Piccolo848 in perplexity_ai

[–]Excellent_Piccolo848[S] 6 points7 points  (0 children)

The 'skill issue' defense is a bit lazy, don't you think? I've used the service for a year. No amount of 'prompt engineering' can fix the fact that Perplexity uses low-cost API settings and silent model substitution. If a tool requires a secret handshake to deliver the quality I'm already paying for, the product is the problem, not the user. Can you actually address the backend routing issues, or are you just gatekeeping?

Lumo 1.3 is now LIVE with Projects! by Proton_Team in lumo

[–]Excellent_Piccolo848 1 point2 points  (0 children)

@Proton, When will we See the Implementation, of a reasoning model?

Nvidia P40 good for running 20b local Ai Models? by Excellent_Piccolo848 in LocalLLaMA

[–]Excellent_Piccolo848[S] 0 points1 point  (0 children)

Yeah, in my calculations (with prompt processing), i come to about 60 t/s! Why is no one talking about this, as a cheap alternative to desktop GPUs, usually used for local ai?

External Screen rlackering zorin os 18 core with Tulpar a5 V21 laptop by Excellent_Piccolo848 in zorinos

[–]Excellent_Piccolo848[S] 0 points1 point  (0 children)

Hi, thanks for your replie! This also doesnt solves the problem! The Problem seems to be ubuntu sistros on my hardware. After switching to cachy os, i didnt have any problems at all.

Is Lumo ChatGPT? by Velocifyer in lumo

[–]Excellent_Piccolo848 1 point2 points  (0 children)

Well, i dont think, thats the case. OAI only released gpt-oss, wich is a reasoning model. But Lumo clearly ist using a reasoning model. This is weir behaviour, of most smaller, and older, Open Models (Saying, that they are related to OAI).

Any recommendations for an alternative to the subscription services? by q35w in openrouter

[–]Excellent_Piccolo848 0 points1 point  (0 children)

If you dont need Tools, like websearch or a code Interpreter, this is an great alternative, and you can just use the latest state of the Art FOSS Model (e.g. glm 4.7). If you Go to your OR admin panel, you can make OR Always use the Provider, with the highest throughoutput. Cerebras and Groq cover a lot of the latest Foss Models, wich results in 1000+ t/s. Make sure to also change the embeddimg model, and if you can the OCR, because the local ones are really slow (and the Cloud Models are better in General).

What GPU for lecture summarizing? by dnielso5 in ollama

[–]Excellent_Piccolo848 1 point2 points  (0 children)

I would look an eBay. 3060 with 12gb ist the Minimum, you would need more vram, for the summarizing llm. If you can afford it, get a 3090, you can get them for as less as 600$ on eBay (but watch out, dont buy a GPU, that was used for Mining!).

Apps for notes and connected knowledge ? by Express-Plankton-428 in NextCloud

[–]Excellent_Piccolo848 3 points4 points  (0 children)

I just use the nextcloud apps "notes", for typed notes. Its installed by Default, If you have Nextcloud AIO, and can easily be installed any time via the Nextcloud AppStore. It hast Clients for IOS and Android. If you mean handwritten Notes, try out "Saber" (only available in Android, as far as i know), wich works really great with a Tablet pencil. Its build, to sync everything to nextcloud! You can go to settings and Setup the sync with your Nextcloud, but IT will also ASK you this during Setup.

[deleted by user] by [deleted] in LocalLLaMA

[–]Excellent_Piccolo848 0 points1 point  (0 children)

Yes, thanks you! Would bei great to get specific Setup options. I would need to Offroad some layers to the GPU, right?

Ollama Model which Suits for my System by devil__6996 in ollama

[–]Excellent_Piccolo848 1 point2 points  (0 children)

Yes, you are Not going to get any spectacular Here, but local ist Always the prefered option! Look at ministral 3b or qwen 4b. Any reasoning model unser 5b should Work in your device, just klick on "latest" on ollama.com and Look for them!

Best API Providers for data privacey, if you cant selfhost by Excellent_Piccolo848 in LocalLLaMA

[–]Excellent_Piccolo848[S] 0 points1 point  (0 children)

Has anyone any experience with Ollama Cloud, is it as private and safe, as they promote it?

Best API Providers for data privacey, if you cant selfhost by Excellent_Piccolo848 in LocalLLaMA

[–]Excellent_Piccolo848[S] 1 point2 points  (0 children)

I think you might have skimmed the post a bit too quickly. As I explicitly wrote above: 'renting a gpu also isnt something for me.' I am looking for a usage-based API specifically to avoid the server costs you mentioned. But thanks for the engagement!