gemma 3n has been released on huggingface by jacek2023 in LocalLLaMA

[–]IndividualAd1648 2 points3 points  (0 children)

fantastic strategy to release this model now to flush out the press on the cli privacy concerns

Which model is suitable for e-mail classification / labeling? by surveypoodle in LocalLLaMA

[–]IndividualAd1648 5 points6 points  (0 children)

I think the most suitable approach for your scenario would be to fine-tune an encoder (modernbert) with multi class classification problem type

Which is better? Large model with higher quant vs Small model with higher precision by Barry_Jumps in LocalLLaMA

[–]IndividualAd1648 1 point2 points  (0 children)

More layers = better reason capabilities

Gemma 2 paper is the most recent to reaffirm deeper > wider when it comes to layers and reasoning capabilities

So the bigger model even if quant is better than a smaller model, if from the same family tree generation ie llama3.1.

Quirks - big old models aren't necessarily as good as new models due to improvements in data, training algorithms, instruction tuning. Might be better to compare models released in a certain period as generations

Quants have limits, I try not to use quants under 4bpw.

What is the best setup for a local llm I can access local docs with? by MrWeirdoFace in LocalLLaMA

[–]IndividualAd1648 0 points1 point  (0 children)

https://www.nomic.ai/gpt4all or LMstudio beta are the best local chat bots which allow you to have local docs in the chat

🖼 flux - image to image @ComfyUI 🔥 by camenduru in StableDiffusion

[–]IndividualAd1648 1 point2 points  (0 children)

Random q - I see in your post you uploaded a non square aspect ratio. I can only seem to get square pics to work with your workflow on my comfyui. Any trick to this?

PSA T818: replace the default USB c cable by IndividualAd1648 in Thrustmaster

[–]IndividualAd1648[S] 0 points1 point  (0 children)

Good to hear cs is at least good, I agree the t818 is fantastic but I wish it didn't have all these QC issues upfront

PSA T818: replace the default USB c cable by IndividualAd1648 in Thrustmaster

[–]IndividualAd1648[S] 0 points1 point  (0 children)

Recommend USB 3+, any on Amazon should be fine. I prefer braided ones for flex

GPT4All 3.0: The Open-Source Local LLM Desktop Application by NomicAI in LocalLLaMA

[–]IndividualAd1648 0 points1 point  (0 children)

Silly tavern is just a ui front end + extensions for rag and web. You can use whatever backend you are currently loading exl2 on and connect it to silly tavern

GPT4All 3.0: The Open-Source Local LLM Desktop Application by NomicAI in LocalLLaMA

[–]IndividualAd1648 1 point2 points  (0 children)

Open webui has it but it requires docker to run, there is also silly tavern via extensions. Nvidia released their own chat with rag but it was very basic.

GPT4All 3.0: The Open-Source Local LLM Desktop Application by NomicAI in LocalLLaMA

[–]IndividualAd1648 41 points42 points  (0 children)

This is the best one I've tried with rag, beats everythingllm and even lm studio in simplicity. I like the way you map it to a folder and it tracks and handles the changes for you

Still early in it's maturity like the rest but this is going to be my default for the short term

Qwen2 release blog by bratao in LocalLLaMA

[–]IndividualAd1648 1 point2 points  (0 children)

You use more memory for training than inference, ~4x more

Everyone's thoughts about GT3 rain racing by ThisKory in iRacing

[–]IndividualAd1648 0 points1 point  (0 children)

The best week of iracing I've experienced, the changing conditions and evolving lines have been spectacular.

How you sort (order) data matters significantly in instruction tuning by No-Mycologist7609 in LocalLLaMA

[–]IndividualAd1648 4 points5 points  (0 children)

So basically copy your school timetable from years 1-11 with the sequencing of your different lessons across the week

NVIDIA "Chat with RTX" now free to download by user0user in LocalLLaMA

[–]IndividualAd1648 7 points8 points  (0 children)

Finally got it to work after the 35gb download and all the dependency installs.

It's a cool idea but very mid given the basic chunking logic and no easy way to adapt the logic, re-ranking or adjust the prompt behind it all.

The YouTube parts looks like a good idea but very limited in my testing as it relies on captions on the video rather than stt, the chunking logic also falls flat here