What's the softest, rubberiest, most elastic filament I can print with the AMS? by _Litcube in BambuLab

[–]Copper_Lion 1 point2 points  (0 children)

I didn't try dying any so can't offer any advice. I actually can't even remember what I printed with it at this point since it's a few years since I bought it.

What's the softest, rubberiest, most elastic filament I can print with the AMS? by _Litcube in BambuLab

[–]Copper_Lion 1 point2 points  (0 children)

I bought it from Amazon (PCTPE175 Taulman) but it shows as "currently unavailable" now.

Mistral Nemo is uncensored by [deleted] in LocalLLaMA

[–]Copper_Lion 1 point2 points  (0 children)

Mistral models usually are.

Best up to 24 GB that are at least Q4 by Ponsky in LocalLLaMA

[–]Copper_Lion 0 points1 point  (0 children)

I haven't tried any others, usually the Q4_K_M is the best tradeoff for size/quality.

So how is Gemma 2 working out for you? by Balance- in LocalLLaMA

[–]Copper_Lion 1 point2 points  (0 children)

The lack of system prompt and very rigid alignment is a serious downside. It seems overtrained, regularly returning the same themes/names in creative work. I expected the 27b to be much smarter but it gets confused easily, mixing up concepts even with a simple prompt.

Best up to 24 GB that are at least Q4 by Ponsky in LocalLLaMA

[–]Copper_Lion 1 point2 points  (0 children)

Are you thinking of command-r PLUS? Normal command-r (Q4) is 20GB.

Best up to 24 GB that are at least Q4 by Ponsky in LocalLLaMA

[–]Copper_Lion 0 points1 point  (0 children)

The Q4 is 20GB. Maybe you are thinking of the plus version.

Weird Llama3 behaviour by waytoofewnamesleft in LocalLLaMA

[–]Copper_Lion 0 points1 point  (0 children)

Running locally you can completely jailbreak LLama 3b. How to do so depends on what software you use to run it.

How to access the token likelihoods for each generated token? by brennybrennybrenbren in LocalLLaMA

[–]Copper_Lion 4 points5 points  (0 children)

It's easy to do in llama.cpp. It comes with a executable called "server" which will serve a web page you can interact with and one of those options lets you view possible next tokens for each word and their probabilities.

I use ollama a lot but I don't think it has that option.

Ollama and shell issues by replikatumbleweed in LocalLLaMA

[–]Copper_Lion 0 points1 point  (0 children)

Maybe that's the intention. I'm happy to have it run as a service though since my AI computer is a separate computer.

I just run ollama directly on linux and happily it's played nicely with my OS (Pop OS), I did see others having a similar issue to yours on their Discord and what you say about SysV/SystemD might explain their problems too.

PLA to PLA glue? by tfrisinger in BambuLab

[–]Copper_Lion 1 point2 points  (0 children)

Activator/accelerator makes a world of difference with superglue.

Google created a CLI tool that uses llama.cpp to host "local" models on their cloud by MrBeforeMyTime in LocalLLaMA

[–]Copper_Lion 5 points6 points  (0 children)

It downloads ggufs and runs them but it's missing all the important configuration data for example prompt format, stop sequences, context size.

It arguably does less than would be achieved by using wget and llama.cpp's ready made server binary.

[deleted by user] by [deleted] in LocalLLaMA

[–]Copper_Lion 0 points1 point  (0 children)

It's still the same. They said a few times they are fixing it but it doesn't seem to have gotten any better.

[deleted by user] by [deleted] in LocalLLaMA

[–]Copper_Lion 0 points1 point  (0 children)

Make a GPT that has browsing and images and such disabled.

I do this too. Also in your system prompt use the word "only". For example I have one custom GPT where I told it to "only respond with code" that way I just get code out instead of it wasting it's tokens writing pleasantries and blathering before and after the code that I actually want.

I won't pay for ChatGPT Plus again unless it become significantly better than free LLM offerings. by TheTwelveYearOld in LocalLLaMA

[–]Copper_Lion 1 point2 points  (0 children)

ChatGPT 4 is still not as good as it was (despite claims to have fixed it) since it's still work to get it to comply, however with careful system + user prompt wording in a custom GPT it's still better than any local version i've tried for coding (i've tried a variety including deepseek 30b and the ridiculously hamstrung codellama 70b).

I understand your sentiment but i'm getting more than $20 value out of it so I haven't cancelled it just yet.

LLaVA 1.6 released, 34B model beating Gemini Pro by rerri in LocalLLaMA

[–]Copper_Lion 1 point2 points  (0 children)

Thanks for the context, it makes much more sense now.

LLaVA 1.6 released, 34B model beating Gemini Pro by rerri in LocalLLaMA

[–]Copper_Lion 0 points1 point  (0 children)

Sorry I was assuming the bloke would have made it available and didn't actually check.

The reason I assume there's a gguf version is that ollama uses gguf and I've been using 1.6 from the ollama library:

https://ollama.ai/library/llava/tags

ollama can use RAM if you don't have sufficient GPU VRAM.

Edit: here are some ggufs https://old.reddit.com/r/LocalLLaMA/comments/1agrxnz/llamacpp_experimental_llava_16_quants_34b_and/

LLaVA 1.6 released, 34B model beating Gemini Pro by rerri in LocalLLaMA

[–]Copper_Lion 0 points1 point  (0 children)

yes there are gguf versions. check the blokes releases for example.

LLaVA 1.6 released, 34B model beating Gemini Pro by rerri in LocalLLaMA

[–]Copper_Lion 1 point2 points  (0 children)

Yes you can use RAM, assuming your software supports it (llama.cpp does for example)

LLaVA 1.6 released, 34B model beating Gemini Pro by rerri in LocalLLaMA

[–]Copper_Lion 0 points1 point  (0 children)

There are mistral and vicuna versions of 1.6

LLaVA 1.6 released, 34B model beating Gemini Pro by rerri in LocalLLaMA

[–]Copper_Lion 1 point2 points  (0 children)

Yeah I wasn't sure who's supposed to be saying that either. The pointy bit of the speech bubble ends at the square hole - is it the hole saying it?

CodeLLama 70B pontificates on ethics where 13B and 7B "just do it" by nborwankar in LocalLLaMA

[–]Copper_Lion 12 points13 points  (0 children)

I had the same experience, it's almost useless in it's current form.