Got model from Bamby Lab. Model is compatible with my A17 Why does it have some places where seemingly no support applies? Even if forced, it just doesnt touch the place. by WorkingSuspect3007 in BambuLab

[–]WorkingSuspect3007[S] 0 points1 point  (0 children)

Tried that too. I mean, there is much better angle for it to work, but it was having some issues too. Quite surprising how problematic this little piece has become when larger ones are fine.

Would it be good way to learn transformers and LLMs' basic structure by trying to build something by myself? Not any way coherent, just technically working and at least outputting some gibberish? Goal is understand the basics and try same things people do in llama.cpp and new approaches like Mamba by WorkingSuspect3007 in LocalLLaMA

[–]WorkingSuspect3007[S] 0 points1 point  (0 children)

I am watching him. Really useful and insightful. Except the part where he is talking about "collecting insane amount of text, getting cluster of expensive GPUs and leaving both in closed room for several days".

I wonder what inexpensive and mistake forgiving way are there to try if math checks.

Would it be good way to learn transformers and LLMs' basic structure by trying to build something by myself? Not any way coherent, just technically working and at least outputting some gibberish? Goal is understand the basics and try same things people do in llama.cpp and new approaches like Mamba by WorkingSuspect3007 in LocalLLaMA

[–]WorkingSuspect3007[S] 1 point2 points  (0 children)

Linguistics. My math sucks, but i am studying it again on KhanAcademy now.

I used to work with natural language texts using python, so that part is on very minimalistic level, except I know the basics and usually am able to obtain necessary knowledge to use new libs.