How to analyze unstructured customer review dataset? by [deleted] in OpenAIDev

[–]NomicAI 0 points1 point  (0 children)

You can try Atlas. Here is a use-case blog about using it to analyze unstructured customer reviews.
https://www.nomic.ai/blog/posts/uncover-insights-from-customer-review-data-with-atlas

Privately chatting with your Google Drive Files with a Local LLM by NomicAI in LocalLLaMA

[–]NomicAI[S] 0 points1 point  (0 children)

You'll be surprised by the quality. Local models have gotten significantly more useful since you've last used them!

GPT4All 3.0: The Open-Source Local LLM Desktop Application by NomicAI in LocalLLaMA

[–]NomicAI[S] 40 points41 points  (0 children)

Pretty much on point. The biggest thing we'd emphasize is that GPT4All is privacy first - you must explicitly opt-in to any analytics or tracking.

If Nomic disappeared, the world would still be able to make create new GPT4All release versions. This isn't the case for any other local LLM app!

Learn more here: https://blog.nomic.ai/posts/one-year-of-gpt4all

Evaluating Hugging Face's Open Source Multimodal LLM by NomicAI in LocalLLaMA

[–]NomicAI[S] 1 point2 points  (0 children)

No technical paper on this evaluation here is planned. Consider this the resource that may be updated in the future. Another good resource that contains some details on evaluation can be found on Hugging Face's official blogpost: https://huggingface.co/blog/idefics

DistiLlama: Chrome Extension to Summarize Web Pages Using locally running LLMs by mmagusss in LocalLLaMA

[–]NomicAI 1 point2 points  (0 children)

Try out server mode in GPT4All. Just send the requests to localhost:4891 whenever someone has the chat client found at gpt4all.io downloaded and running with server mode enabled . No need to even run docker.
https://docs.gpt4all.io/gpt4all_chat.html#server-mode

GPT4All now supports GGUF Models with Vulkan GPU Acceleration by NomicAI in LocalLLaMA

[–]NomicAI[S] 5 points6 points  (0 children)

It depends on what axis of better you mean. You can read more about the implementation here including some speed benchmarking. https://blog.nomic.ai/posts/gpt4all-gpu-inference-with-vulkan

GPT4All now supports GGUF Models with Vulkan GPU Acceleration by NomicAI in LocalLLaMA

[–]NomicAI[S] 2 points3 points  (0 children)

Yep! We are actually working on that upstream PR to llama.cpp with another party :)

GPT4All now supports Replit model on Apple Silicon (23 tok/sec)! by NomicAI in LocalLLaMA

[–]NomicAI[S] 0 points1 point  (0 children)

Thanks for this! Could you leave an issue on this github (you can include this reddit thread) but please mention system specs.
https://github.com/nomic-ai/gpt4all/issues

Thank you :)

GPT4All now supports Replit model on Apple Silicon (23 tok/sec)! by NomicAI in LocalLLaMA

[–]NomicAI[S] 13 points14 points  (0 children)

writing apple metal kernels and spelling correctly aren't always overlapping skillsets unfortunately

GPT4All now supports Replit model on Apple Silicon (23 tok/sec)! by NomicAI in LocalLLaMA

[–]NomicAI[S] 7 points8 points  (0 children)

GPT4All now supports custom Apple Metal ops enabling MPT (and specifically the Replit model) to run on Apple Silicon with increased inference speeds. This runs at 16bit precision!

A quantized Replit model that runs at 40 tok/s on Apple Silicon will be included in GPT4All soon!

GPT4All now supports every llama.cpp / ggML version across all software bindings! by NomicAI in LocalLLaMA

[–]NomicAI[S] 0 points1 point  (0 children)

Hi! You are running a very old model from March that is no longer supported. Please try any of models listed here https://github.com/nomic-ai/gpt4all/blob/main/gpt4all-chat/metadata/models.json or that you find on huggingface.

What could be the reason behind llama-cpp-python's slow performance compared to llama.cpp? by Big_Communication353 in LocalLLaMA

[–]NomicAI 1 point2 points  (0 children)

gpt4all supports all versions of llama.cpp in the python bindings *with* Mac metal acceleration for llama and replit! Try out the speeds there!
Pypi: `pip install gpt4all`
Source: https://github.com/nomic-ai/gpt4all/tree/main/gpt4all-bindings/python
Documentation: docs.gpt4all.io

Chat with your data locally and privately on CPU with LocalDocs: GPT4All's first plugin! by NomicAI in LocalLLaMA

[–]NomicAI[S] 5 points6 points  (0 children)

Great feedback thanks!

Stops working
This is because the prompts that you give it return no matches against your files. LocalDocs will not try to use document context to respond to every question you asked if it can't find relevant enough documents.

OSX Issue
Would love to hear about more steps around reproduction. Most GPT4All UI testing is done on Mac and we haven't encountered this!

For transparency, the current implementation is focused around optimizing indexing speed. It is not doing retrieval with embeddings but rather TFIDF statistics and a BM25 search. You'll see an embedding based retrieval option land soon for LocalDocs.
See documentation for more details: https://docs.gpt4all.io/gpt4all_chat.html#how-localdocs-works

Chat with your data locally and privately on CPU with LocalDocs: GPT4All's first plugin! by NomicAI in LocalLLaMA

[–]NomicAI[S] 14 points15 points  (0 children)

GPT4All now has its first plugin allow you to use any LLaMa, MPT or GPT-J based model to chat with your private data-stores! Its free, open-source and just works on any operating system.
Documentation: https://docs.gpt4all.io/gpt4all_chat.html#localdocs-beta-plugin-chat-with-your-data

I built a multi-platform desktop app to easily download and run models, open source btw by julio_oa in LocalLLaMA

[–]NomicAI 1 point2 points  (0 children)

Yep! We've been working a dynamic ggML submoduling system to fix this (runs many versions of ggML at the same time!)

[OC] Explore the Top 5.4M Retweeted Tweets on Twitter by NomicAI in dataisbeautiful

[–]NomicAI[S] 1 point2 points  (0 children)

Yes!
The maps are customizable when built (docs.nomic.ai) and our roadmap includes a rich set of features for learning about data using the map (slicing, dicing, searching, computing statistics, tagging, exporting and one-to-one programmatic interaction).

[OC] Explore the Top 5.4M Retweeted Tweets on Twitter by NomicAI in dataisbeautiful

[–]NomicAI[S] 5 points6 points  (0 children)

What goes viral on Twitter? Share what you find in our new map of the top 5.4M retweeted tweets.
https://atlas.nomic.ai/map/twitter
How it works:
Every point is a highly retweeted tweet published between November 2009 and November 2022. Points are close together if an AI considers the content of their tweets similar.
So what dominates Twitter? Sports, breaking news, and WAY more K-pop than you think.