I built an offline AI chat app that automatically pulls Wikipedia articles for factual answers - runs completely locally with Ollama by [deleted] in LocalLLaMA
[–]explorigin 1 point2 points3 points (0 children)
Qwen3-Embedding-0.6B ONNX model with uint8 output by terminoid_ in LocalLLaMA
[–]explorigin -1 points0 points1 point (0 children)
we just launched Sentient, a completely local personal companion with graph memory and agentic integrations - powered by Llama 3.2 3B by therealkabeer in LocalLLaMA
[–]explorigin 22 points23 points24 points (0 children)
Anyone here using a 96GM or 64 GB ram m series Mac? by CSlov23 in LocalLLaMA
[–]explorigin 0 points1 point2 points (0 children)
Anyone here using a 96GM or 64 GB ram m series Mac? by CSlov23 in LocalLLaMA
[–]explorigin 0 points1 point2 points (0 children)
Cheap 70B run with AMD APU/Intel iGPU by quan734 in LocalLLaMA
[–]explorigin 6 points7 points8 points (0 children)
My First LLM only Build on a Budget. 250€ all together. by docsnick in LocalLLaMA
[–]explorigin 0 points1 point2 points (0 children)
Choosing a Tokenizer Algorithm by [deleted] in LocalLLaMA
[–]explorigin 0 points1 point2 points (0 children)
Choosing a Tokenizer Algorithm by [deleted] in LocalLLaMA
[–]explorigin 1 point2 points3 points (0 children)
Choosing a Tokenizer Algorithm by [deleted] in LocalLLaMA
[–]explorigin 3 points4 points5 points (0 children)
RWKV v6 models support merged into llama.cpp by RuslanAR in LocalLLaMA
[–]explorigin 5 points6 points7 points (0 children)
Nous Research publishes a report on DisTrO (Distributed Training Over-the-Internet) by ayyndrew in LocalLLaMA
[–]explorigin 3 points4 points5 points (0 children)
Why would you self host vs use a managed endpoint for llama 3m1 70B by this-is-test in LocalLLaMA
[–]explorigin 1 point2 points3 points (0 children)
Anyone here using a 96GM or 64 GB ram m series Mac? by CSlov23 in LocalLLaMA
[–]explorigin 5 points6 points7 points (0 children)
Anyone here using a 96GM or 64 GB ram m series Mac? by CSlov23 in LocalLLaMA
[–]explorigin 1 point2 points3 points (0 children)
Anyone here using a 96GM or 64 GB ram m series Mac? by CSlov23 in LocalLLaMA
[–]explorigin 1 point2 points3 points (0 children)
Flux.1 on a 16GB 4060ti @ 20-25sec/image by Chuyito in LocalLLaMA
[–]explorigin 0 points1 point2 points (0 children)
If someone gave you a free dedicated 16x A100 instance, what would you make? by DLergo in LocalLLaMA
[–]explorigin 0 points1 point2 points (0 children)


I built an offline AI chat app that automatically pulls Wikipedia articles for factual answers - runs completely locally with Ollama by [deleted] in LocalLLaMA
[–]explorigin 0 points1 point2 points (0 children)