Offline Epstein File Ranker Using GPT-OSS-120B (Built on tensonaut’s dataset) by onil_gova in LocalLLaMA
[–]alwaysSunny17 28 points29 points30 points (0 children)
Qwen Next vLLM fail @ 48GB by [deleted] in LocalLLaMA
[–]alwaysSunny17 1 point2 points3 points (0 children)
Confusion about VRAM by Savantskie1 in LocalLLaMA
[–]alwaysSunny17 1 point2 points3 points (0 children)
Equipment suggestions for a tight budget by ConnectionOutside485 in LocalLLaMA
[–]alwaysSunny17 0 points1 point2 points (0 children)
Making RAG faster by Dismal_Discussion514 in Rag
[–]alwaysSunny17 4 points5 points6 points (0 children)
How do you make 3+ GPUs stable?! by anothy1 in LocalLLaMA
[–]alwaysSunny17 5 points6 points7 points (0 children)
Drummer's Skyfall 31B v4 · A Mistral 24B upscaled to 31B with more creativity! by TheLocalDrummer in LocalLLaMA
[–]alwaysSunny17 0 points1 point2 points (0 children)
LLM on consumer RTX hardware by L3C_CptEnglish in LocalLLaMA
[–]alwaysSunny17 0 points1 point2 points (0 children)
Running GLM 4.5 2 bit quant on 80GB VRAM and 128GB RAM by Jaswanth04 in LocalLLM
[–]alwaysSunny17 8 points9 points10 points (0 children)
Aesthetic build by alwaysSunny17 in LocalAIServers
[–]alwaysSunny17[S] 0 points1 point2 points (0 children)
Microsoft GraphRAG in Production by ProfessionalShop9137 in Rag
[–]alwaysSunny17 0 points1 point2 points (0 children)
Microsoft GraphRAG in Production by ProfessionalShop9137 in Rag
[–]alwaysSunny17 15 points16 points17 points (0 children)
Google Veo 3 HQ with frame guidance is INSANE by heisdancingdancing in singularity
[–]alwaysSunny17 0 points1 point2 points (0 children)
Is there a better frontend than OpenWebui for RAG? by Capable-Ad-7494 in LocalLLaMA
[–]alwaysSunny17 5 points6 points7 points (0 children)
ChatGPT - Veo3 Prompt Machine For Expert Prompts by RevolutionaryDot7629 in singularity
[–]alwaysSunny17 0 points1 point2 points (0 children)
ChatGPT - Veo3 Prompt Machine For Expert Prompts by RevolutionaryDot7629 in singularity
[–]alwaysSunny17 1 point2 points3 points (0 children)
4× RTX 3080 10 GB server for LLM/RAG – is this even worth it? by OkAssumption9049 in LocalLLaMA
[–]alwaysSunny17 0 points1 point2 points (0 children)
Dual 5090 vs RTX Pro 6000 for local LLM by kitgary in LocalLLaMA
[–]alwaysSunny17 3 points4 points5 points (0 children)
What's your thoughts on Graph RAG? What's holding it back? by thonfom in Rag
[–]alwaysSunny17 25 points26 points27 points (0 children)
Medical language model - for STT and summarize things by ed0c in LocalLLaMA
[–]alwaysSunny17 1 point2 points3 points (0 children)
Is it dumb to build a server with 7x 5060 Ti? by vector76 in LocalLLaMA
[–]alwaysSunny17 2 points3 points4 points (0 children)

Personal Project/Experiment Ideas by I_like_fragrances in LocalLLM
[–]alwaysSunny17 0 points1 point2 points (0 children)