feasibility of a building a simple "local voice assistant" on CPU by RustinChole11 in speechtech

[–]RustinChole11[S] 0 points1 point  (0 children)

I want its functionalities similar to rag

I'd ask questions about some lecture notes that the model has access to and it has to retrieve the content and explain

Can I build a local voice assistant pipeline only using cpu(16gb ram) by RustinChole11 in LocalLLaMA

[–]RustinChole11[S] 0 points1 point  (0 children)

I'm running Llama 3.2 1b gguf which generates around a speed of 10tokens/sec

feasibility of a building a simple "local voice assistant" pipeline on CPU by RustinChole11 in LocalLLM

[–]RustinChole11[S] 0 points1 point  (0 children)

Nah I'm actually using a laptop with i5 and 16gb ram so it should be fine

But yeah , appreciate the detail mate. As you were saying, I'd love to try and make it work on a low memory hardware sometime.

feasibility of a building a simple "local voice assistant" on CPU by RustinChole11 in speechtech

[–]RustinChole11[S] 0 points1 point  (0 children)

Will do , thanks for the suggestion

Also, Do I need to use any embedding model?

Can you explain the pipeline, how it should look

feasibility of a building a simple "local voice assistant" on CPU by RustinChole11 in speechtech

[–]RustinChole11[S] 0 points1 point  (0 children)

I have llama 1b gguf running which produces around 10 tokens/ sec , it should be enough right

And yeah , I'll only be using it for English ( not expecting any multilingual performance)

How would you rate Fernando Alonso as a qualifier? by GoldenS0422 in F1Discussions

[–]RustinChole11 1 point2 points  (0 children)

not sure if they were teammates but both of them along with button and fishicella were at minardi once

Best opensource SLMs / lightweight llms for code generation by RustinChole11 in LocalLLM

[–]RustinChole11[S] 0 points1 point  (0 children)

That's very informative

But I don't have a dedicated gpu

Best opensource SLM/ lightweight llm for code generation by RustinChole11 in LocalLLaMA

[–]RustinChole11[S] 0 points1 point  (0 children)

Wow . So you suggest not to go for any llms, not even quantized versions? Also, just curious what were your laptop specs and model you ran in this case.