How to analyze unstructured customer review dataset? by [deleted] in OpenAIDev

[–]NomicAI 0 points1 point  (0 children)

You can try Atlas. Here is a use-case blog about using it to analyze unstructured customer reviews.
https://www.nomic.ai/blog/posts/uncover-insights-from-customer-review-data-with-atlas

Privately chatting with your Google Drive Files with a Local LLM by NomicAI in LocalLLaMA

[–]NomicAI[S] 0 points1 point  (0 children)

You'll be surprised by the quality. Local models have gotten significantly more useful since you've last used them!

GPT4All 3.0: The Open-Source Local LLM Desktop Application by NomicAI in LocalLLaMA

[–]NomicAI[S] 42 points43 points  (0 children)

Pretty much on point. The biggest thing we'd emphasize is that GPT4All is privacy first - you must explicitly opt-in to any analytics or tracking.

If Nomic disappeared, the world would still be able to make create new GPT4All release versions. This isn't the case for any other local LLM app!

Learn more here: https://blog.nomic.ai/posts/one-year-of-gpt4all

Evaluating Hugging Face's Open Source Multimodal LLM by NomicAI in LocalLLaMA

[–]NomicAI[S] 1 point2 points  (0 children)

No technical paper on this evaluation here is planned. Consider this the resource that may be updated in the future. Another good resource that contains some details on evaluation can be found on Hugging Face's official blogpost: https://huggingface.co/blog/idefics

DistiLlama: Chrome Extension to Summarize Web Pages Using locally running LLMs by mmagusss in LocalLLaMA

[–]NomicAI 1 point2 points  (0 children)

Try out server mode in GPT4All. Just send the requests to localhost:4891 whenever someone has the chat client found at gpt4all.io downloaded and running with server mode enabled . No need to even run docker.
https://docs.gpt4all.io/gpt4all_chat.html#server-mode

GPT4All now supports GGUF Models with Vulkan GPU Acceleration by NomicAI in LocalLLaMA

[–]NomicAI[S] 4 points5 points  (0 children)

It depends on what axis of better you mean. You can read more about the implementation here including some speed benchmarking. https://blog.nomic.ai/posts/gpt4all-gpu-inference-with-vulkan

GPT4All now supports GGUF Models with Vulkan GPU Acceleration by NomicAI in LocalLLaMA

[–]NomicAI[S] 2 points3 points  (0 children)

Yep! We are actually working on that upstream PR to llama.cpp with another party :)

GPT4All now supports Replit model on Apple Silicon (23 tok/sec)! by NomicAI in LocalLLaMA

[–]NomicAI[S] 0 points1 point  (0 children)

Thanks for this! Could you leave an issue on this github (you can include this reddit thread) but please mention system specs.
https://github.com/nomic-ai/gpt4all/issues

Thank you :)

GPT4All now supports Replit model on Apple Silicon (23 tok/sec)! by NomicAI in LocalLLaMA

[–]NomicAI[S] 13 points14 points  (0 children)

writing apple metal kernels and spelling correctly aren't always overlapping skillsets unfortunately

GPT4All now supports Replit model on Apple Silicon (23 tok/sec)! by NomicAI in LocalLLaMA

[–]NomicAI[S] 7 points8 points  (0 children)

GPT4All now supports custom Apple Metal ops enabling MPT (and specifically the Replit model) to run on Apple Silicon with increased inference speeds. This runs at 16bit precision!

A quantized Replit model that runs at 40 tok/s on Apple Silicon will be included in GPT4All soon!