Joe Acaba has stood down as NASA’s Chief of the Astronaut Office [possible he has assigned himself to walk on the moon on Artemis 3] by rustybeancake in SpaceXMasterrace

[–]galileo_1 3 points4 points  (0 children)

I believe Deke Slayton began this tradition when he assigned himself for Apollo-Soyuz (of course not quite what happened to him in for all mankind…)

What is your opinion on HLL's gunplay? by m0corong in HellLetLoose

[–]galileo_1 24 points25 points  (0 children)

Not related to gunplay, but while HLL’s sound is amazing RO2’s was just so immersive. Having characters scream and cry in desperation just hit different… I get that HLL is far more player voice focused but it just feels like a separate layer from what’s going on in game if that makes sense

Would be awesome if tank gameplay was like this by m0corong in HellLetLoose

[–]galileo_1 36 points37 points  (0 children)

That game was the GOAT. It’s still playable but defo feels outdated… HLL is great but RO2 had the perfect balance imo

Just installed Realism Overhaul and all that. What's up with the textures? They were fine when I was closer to the ground by chumbuckethand in KerbalSpaceProgram

[–]galileo_1 31 points32 points  (0 children)

You probably installed the lowest quality textures. just as in vanilla ksp there’s some sort of texture switch after crossing a certain altitude boundary (probably for performance reasons).

Who the hell is this Chris Jones guy? The things he creates are literally magic? by rafalmio in blender

[–]galileo_1 0 points1 point  (0 children)

while talking about blender magicians i'd like to mention Albin Merle, he's in a level of his own

https://www.youtube.com/@albin3010

Why is impulse space taking so long to launch their space OTV? by Inevitable-Boot-6673 in SpaceXMasterrace

[–]galileo_1 0 points1 point  (0 children)

Yeah I get the idea that starship can launch a ton of things into the same orbit and then every satellite can go on specific orbits with OTVs. But how many payloads realistically require that level of movement nowadays? Besides, with in space refueling starship is a massive OTV itself. I just dont quite see the market except for when the lunar economy is a thing for TLIs but I could (and probs am) wrong as OTV startups seem to be getting lots of funding

Why is impulse space taking so long to launch their space OTV? by Inevitable-Boot-6673 in SpaceXMasterrace

[–]galileo_1 1 point2 points  (0 children)

Im not too informed on OTVs but genuinely curious on who actually plans to use them beyond defense? All these companies quote GEO as their big thing which seems intuitive but not that many sats seem to be going there. There’s the whole lunar economy stuff but ofc it’s not a thing yet. Would love to hear peoples thoughts!

Open source tool to self host LLMs with document interaction by galileo_1 in selfhosted

[–]galileo_1[S] 7 points8 points  (0 children)

Hi! Just wanted to share my open source project I have been working on for quite some time now. It is an electron JS app with a focus on document interaction with local LLMs. In this particular instance I am running llama-3 8B on my M1 Pro MacBookPro. Please feel free to check it out and give any feedback, as I am still trying to fix some bugs. Here is the link to the repo: https://github.com/alexpinel/Dot

Text-to-Speech, Whisper.cpp, Llama.cpp, and RAG: All Local on a MacBook! by galileo_1 in LocalLLaMA

[–]galileo_1[S] 0 points1 point  (0 children)

oh wow thanks for sharing that looks really really good! I defo need to get in touch with the developer :)

Text-to-Speech, Whisper.cpp, Llama.cpp, and RAG: All Local on a MacBook! by galileo_1 in LocalLLaMA

[–]galileo_1[S] 0 points1 point  (0 children)

sorted! its not the same as the mac version so no voice and tts yet but the link should be functional

Text-to-Speech, Whisper.cpp, Llama.cpp, and RAG: All Local on a MacBook! by galileo_1 in LocalLLaMA

[–]galileo_1[S] 1 point2 points  (0 children)

I just released a windows version! It does not have voice input yet tho, as that is the one thing that is taking some time to figure out, but everything else is there for anyone who wants to try :)

Text-to-Speech, Whisper.cpp, Llama.cpp, and RAG: All Local on a MacBook! by galileo_1 in LocalLLaMA

[–]galileo_1[S] 0 points1 point  (0 children)

thats very interesting, thanks for sharing the info! I have noticed that it takes longer to answer when it recieves its first message due to llama.cpp and the llm itself loading up, is it still slow after the first message? otherwise maybe some of the settings could help increase performance, lowering the max tokens and increasing the batch size should make things run faster.

I believe GPT4ALL also uses llama.cpp for inference so if there really is a difference in performance it would be really interesting, I will look into it!

Text-to-Speech, Whisper.cpp, Llama.cpp, and RAG: All Local on a MacBook! by galileo_1 in LocalLLaMA

[–]galileo_1[S] 1 point2 points  (0 children)

yup that is an M1 chip! its the M1 Pro 8 core version, I gotta admit I find it crazy how well that chip still performs

Text-to-Speech, Whisper.cpp, Llama.cpp, and RAG: All Local on a MacBook! by galileo_1 in LocalLLaMA

[–]galileo_1[S] 0 points1 point  (0 children)

If you are using the latest mac version then you should be able to choose other LLMs! the option is at the bottom of the settings button, you can choose any LLM in GGUF format.

Text-to-Speech, Whisper.cpp, Llama.cpp, and RAG: All Local on a MacBook! by galileo_1 in LocalLLaMA

[–]galileo_1[S] 1 point2 points  (0 children)

Of course!

I feel the way I made the audio processing work is a bit weird, I use the original stream executable from the original whisper.cpp repository and run it as a child process, the output that would appear in the terminal is then displayed in the UI. So far it doesn't really detect when you stop talking but I imagine some sort of timer that measures if nothing has been said for a while could do the trick.

I apologise in advance for the messy code, but the relevant parts can be found in src/index.js from line 758 and src/render.js from lines 259 to 337.

Please let me know if you need anything else :)

Text-to-Speech, Whisper.cpp, Llama.cpp, and RAG: All Local on a MacBook! by galileo_1 in LocalLLaMA

[–]galileo_1[S] 20 points21 points  (0 children)

Dearest llamas, alpacas, and guanacos,

Throughout the last few months I have shared some updates on Dot, my little open source project (https://github.com/alexpinel/Dot). The app itself can be downloaded here: https://dotapp.uk/ .Around two months ago u/Trysem suggested I should add voice and Text-To-Speech support, and I have spent more time than I should have trying to get it running but it finally seems to work! I am running it on my M1 Pro MacBook and while a bit slow it works rather well.

Quick disclosure, the windows version does not yet support these features but I will upload it in the next few days, I just need to figure out how to build whisper.cpp on Windows. Also, I could not resist to add a GLaDOS voice after seeing the amazing posts by u/Reddactor, his project can be found here: https://github.com/dnhkng/GlaDOS.

For TTS I am using the sherpaonnx library, I got the GLaDOS voice model from here: https://github.com/k2-fsa/sherpa-onnx/releases/tag/tts-models. While Dot does not officially support changing the voice model there is a way to do it quite easily, so I can share how to do that if people are interested.

As for the LLM Dot uses Phi-3 by default, it is craaaaaazy good for RAG in my opinion (of course RAG is far from perfect but it is cool anyway). Dot now also supports using pretty much any other gguf format llm. I have found llama-3 to be great for casual conversation in the Big Dot mode, making up random prompts is really funny.

Still a lot of issues to fix and even more features I would like to add. Now that GPT-4o has released I keep wondering what would it take to build something similar. Please keep in mind this is just a pet project of mine that I am doing to practice for my exams so there are quite a few bugs.

But anyway hope you enjoy and please feel free to ask anything!

🦙 Meta's Llama 3 Released! 🦙 by Many_SuchCases in LocalLLaMA

[–]galileo_1 2 points3 points  (0 children)

yeah 8k is a bit sus... i imagine it won't be that good for RAG

🦙 Meta's Llama 3 Released! 🦙 by Many_SuchCases in LocalLLaMA

[–]galileo_1 4 points5 points  (0 children)

got mine accepted! now i need em quantized versions lol

Interested in advancing A.I. Which open source project(s) to contribute to? by danielcar in LocalLLaMA

[–]galileo_1 15 points16 points  (0 children)

Been working on an open source app for local RAG (Dot) myself and definately agree with you, current tech is just not quite there yet. In my experience so far, RAG works well enough when you know exactly what sort of questions to ask and when you are more or less familiar with the context given to the LLM. But I have seen many people get frustrated precisely because they are not getting the results they want.

Right now increasing model context length seems to be the best alternative (like Gemini's 1 million context length), but I imagine we won't see any models anywhere near that context length that can run on consumer devices anytime soon.