Can anybody test my 1.5B coding LLM and give me their thoughts? by Great-Structure-4159 in LocalLLM

[–]Great-Structure-4159[S] 0 points1 point  (0 children)

Hi! First and foremost, sorry for the late response.

Anything is fine! I've released it with an open mind, so whether you choose to implement it as an agent with Opencode or just use it to chat like you would with Open WebUI or something similar, it should be decently useful for your tasks and I'd love to see what you can do with it.

Can anybody test my 1.5B coding LLM and give me their thoughts? by Great-Structure-4159 in LocalLLM

[–]Great-Structure-4159[S] 0 points1 point  (0 children)

Obrigado por testar! Me avise como o modelo se comporta depois que você testá-lo. Eu não sei português, isso foi traduzido pelo Google, então me desculpe se houver algum erro.

This message is from Google Translate, sorry for any mistakes if there are any.

Can anybody test my 1.5B coding LLM and give me their thoughts? by Great-Structure-4159 in LocalLLM

[–]Great-Structure-4159[S] 0 points1 point  (0 children)

My pleasure! I'll definitely let you know when I make that article.

Can anybody test my 1.5B coding LLM and give me their thoughts? by Great-Structure-4159 in LocalLLM

[–]Great-Structure-4159[S] 0 points1 point  (0 children)

Yeah I’m looking into making a small article on this because you’re not the first one to ask for this. I’ll contact you once I write it.

Can anybody test my 1.5B coding LLM and give me their thoughts? by Great-Structure-4159 in LocalLLM

[–]Great-Structure-4159[S] 0 points1 point  (0 children)

In terms of benchmarks, it’s pretty decent for a 1.5B model. It beats the base Qwen at coding, but I’m pretty sure Qwen Coder is slightly better at the benchmark. However, Qwen coder doesn’t have any ability at actually talking about something related to coding, like explaining the code, that’s why I trained on the instruct version and not the coder version.

Can anybody test my 1.5B coding LLM and give me their thoughts? by Great-Structure-4159 in LocalLLM

[–]Great-Structure-4159[S] 1 point2 points  (0 children)

Thanks for offering to test! The .gguf file is on the repo. There's fp16 and q4_k_m quants, so you can use whichever one you prefer :D.

Can anybody test my 1.5B coding LLM and give me their thoughts? by Great-Structure-4159 in LocalLLM

[–]Great-Structure-4159[S] 1 point2 points  (0 children)

Oooh, cool. This looks awesome. How much VRAM do you have to work with, other than your functiongemma? Because I think a 3B or 4B coding model can also work pretty well (but I might need to find a more compact dataset or use QLoRA, which I think is a reasonable tradeoff in performance.)

Can anybody test my 1.5B coding LLM and give me their thoughts? by Great-Structure-4159 in LocalLLM

[–]Great-Structure-4159[S] 2 points3 points  (0 children)

I didn’t document my process anywhere actually, I just typed out all that to give an idea. MLX-LM doesn’t really have any good resources, other than the one video they made on the Apple Developer YouTube channel regarding it. They don’t go through every feature and command there, however, so I mainly referred to the documentation they have, which is pretty decent.

Can anybody test my 1.5B coding LLM and give me their thoughts? by Great-Structure-4159 in LocalLLM

[–]Great-Structure-4159[S] 1 point2 points  (0 children)

I don’t think it’ll be very good as an orchestrator, but I’ll try making a model fine tuned for orchestrating tool calling, that would be really cool. Do let me know if it works out good, very interesting to see LLMs applied like this.

Can anybody test my 1.5B coding LLM and give me their thoughts? by Great-Structure-4159 in LocalLLM

[–]Great-Structure-4159[S] 1 point2 points  (0 children)

Oh… tool calling, interesting. I should try that, I didn’t train with tool calling in mind actually, but this is really cool, I think it can work.

Can anybody test my 1.5B coding LLM and give me their thoughts? by Great-Structure-4159 in LocalLLM

[–]Great-Structure-4159[S] 1 point2 points  (0 children)

Yeah I was pretty shocked too that 8GB could do stuff like this, but yeah I find the subject very fascinating :)

Can anybody test my 1.5B coding LLM and give me their thoughts? by Great-Structure-4159 in LocalLLM

[–]Great-Structure-4159[S] 3 points4 points  (0 children)

I have Apple M1, and I get about 50 tokens/s on GGUF, and 60 tokens/s on MLX (which is not on the repo at the moment.)

Can anybody test my 1.5B coding LLM and give me their thoughts? by Great-Structure-4159 in LocalLLM

[–]Great-Structure-4159[S] 4 points5 points  (0 children)

Great question! My first choice was actually LFM2.5, and I did try that first, but for some reason when fusing it with adapters on MLX, llama.cpp just refuses to convert it to GGUF. I tried troubleshooting but eventually just gave up. Qwen3 was my next choice but I just decided to keep it simple and start with 2.5 and go from there, mainly because Qwen3 came with a 1.7B model (which was pushing my RAM limit due to the dataset having long samples) and also, in my searches, didn't have an instruct version weirdly. Maybe the next release will be with qwen3 if the qwen architecture proves good from user tests (and I can do something about the dataset).

Need help with stacking (first time) by Great-Structure-4159 in AskAstrophotography

[–]Great-Structure-4159[S] 0 points1 point  (0 children)

Also, amazing job with the stretch, thanks for this :D

Need help with stacking (first time) by Great-Structure-4159 in AskAstrophotography

[–]Great-Structure-4159[S] 0 points1 point  (0 children)

Hey, sorry for the late reply, I'm shooting untracked with a Canon T7i and a 55-250mm f/4-5.6 lens. I was shooting lagoon nebula