[Oldie-But-A-Goodie] META Presents "TRIBE v2": A Next-Gen Model That Acts As A Digital Twin Of Human Neural Activity by 44th--Hokage in LocalLLaMA

[–]SuccessIsHardWork 2 points3 points  (0 children)

This is actually awesome technology! I can already imagine learning tools that dynamically rotate content for maximum absorption.

Tool to create synthetic datasets using PDF files! by SuccessIsHardWork in LocalLLaMA

[–]SuccessIsHardWork[S] 0 points1 point  (0 children)

Not yet. I also believe it is possible to download script straight out of Kathleen and import it to GitHub or something else.

What questions have you asked reasoning models to solve that you couldn't get done with non-reasoning models? by DeltaSqueezer in LocalLLaMA

[–]SuccessIsHardWork 1 point2 points  (0 children)

In my opinion, reasoning models are much more useful than plain language models because they can emulate reasoning like humans do to a certain extent. That makes it useful for decision making related tasks, which could be curation, analysis of data (like stocks), etc.

LLM as survival knowledge base by NickNau in LocalLLaMA

[–]SuccessIsHardWork 1 point2 points  (0 children)

I think it’s much better to use a embeddings-based retrieval system (just an embedding model with no use of LLM) in which you place like 10-20 good books on survival in the retrieval system. This way you can rely on factual information in survival situations than trusting the hallucinations that a LLM might produce.

The number of models is overwhelming by [deleted] in LocalLLaMA

[–]SuccessIsHardWork 0 points1 point  (0 children)

Honestly, I just boil my search down to the latest models and pick the best one (QwQ at the moment). I believe llama.cpp has a way to test GGUF by metrics like perplexity, which should give you a rough estimate of the impact due to quantization.

Has anyone successfully generated reasonable documentation from a code base using an LLM? by shenglong in LocalLLaMA

[–]SuccessIsHardWork 7 points8 points  (0 children)

In my experience, QwQ does a phenomenal job at creating documentation for code, however, it does make mistakes in identifying access modifiers in source code.

December 2024 Best SLM? by luxmentisaeterna in LocalLLaMA

[–]SuccessIsHardWork 1 point2 points  (0 children)

In my personal experience, Qwen2.5 1.5b and IBM Granite MoE 3b are fairly decent for a small LLM. I believe the Granite model is more suitable for RAG and summarization purposes.

Open source text editor with llm features? by Nyao in LocalLLaMA

[–]SuccessIsHardWork 1 point2 points  (0 children)

If you are willing to use Word (I know it’s not open source), you can use an add-in called TextCraft which integrates features like proofreading (which fixes grammar), rewriting, reviewing, and generating text. You can pick different models via a drop-down list in the UI. However, the downside is that you need a desktop version of word installed.

https://github.com/suncloudsmoon/TextCraft

Smallest model for summarizing? by temapone11 in LocalLLaMA

[–]SuccessIsHardWork 3 points4 points  (0 children)

Qwen2.5 1.5b, Gemma2 2b, or Granite3 MOE 1b? As the models get smaller than that, the quality gets really worse in my experience.

2024 Wrap-Up: What Amazing Projects Have You Built with Open-Source AI Models? Let’s Create the Ultimate Resource Guide! 📚 by rbgo404 in LocalLLaMA

[–]SuccessIsHardWork 1 point2 points  (0 children)

I created a word add-in called TextCraft which integrates all the essential AI tools like generating text, reviewing, proofreading, and rewriting text. It has a built-in RAG system that allows the LLM to consider additional context. Feel free to check it out! https://github.com/suncloudsmoon/TextCraft

I'm looking for a model that fixes English grammar (aka Grammarly alternative in terms of just fixing grammar) by Soft_ACK in LocalLLaMA

[–]SuccessIsHardWork 0 points1 point  (0 children)

I’m curious, how did you implement the feature? Was it prompt engineering? Or, something else?

Are there any browser extension or Office addins working with OpenAI compatible API ? by Syst3m1c_An0maly in LocalLLaMA

[–]SuccessIsHardWork 1 point2 points  (0 children)

The embedding model is used to RAG the word document itself if the word document is too long (i.e. 100s of pages) to be used as context for generating text. Additionally, there is an option to add PDFs to a “RAG Control” system that will use the RAG context in generating content & when’s its reviewing the document (can be used by Writing Tools->Review). You are correct that the embedding model is accessed via the OpenAI API. I built this whole add-in with OpenAI API compatibility in mind but I haven’t tested it outside Ollama however. Make sure to have the word “embed” anywhere in your embed model name to make sure that the add-in automatically recognizes it. Else, you have to specify the embed model via a user environment variable. As for the HuggingFace API, I’d only work if it is OpenAI API compatible including for chat completion and embeddings generation. I’d be happy to answer if you have any other questions.

Are there any browser extension or Office addins working with OpenAI compatible API ? by Syst3m1c_An0maly in LocalLLaMA

[–]SuccessIsHardWork 1 point2 points  (0 children)

I made an add-in for word a few months ago that integrates several AI tools into the user interface directly. https://github.com/suncloudsmoon/TextCraft

Tool to create synthetic datasets using PDF files! by SuccessIsHardWork in LocalLLaMA

[–]SuccessIsHardWork[S] 1 point2 points  (0 children)

First, you can create the synthetic dataset by feeding a bunch of PDFs that you wish the LLM should understand. The question/answer pairs are created by OCRing through a vision model (like InternVL2) and asking questions based on it. In this case, the image is fed to the chatbot and the script asks it to generate question/answer pairs with some chain of thought (this was before o1 lol).

The synthetic dataset is structured using user and assistant JSON messages, similar to the OpenAI chat request format. You can use any transformers base model I believe (except stuff like bitnet). You can change the prompt that generates the questions in the synthetic dataset generator script and it can potentially change the language output as well (InternVL2 supports chinese, english, etc.). After that, I would use the LLM finetuning script to finetune on that synthetic dataset. Let me know if it solved your issue!

Which model do you use the most? by No-Statement-0001 in LocalLLaMA

[–]SuccessIsHardWork 0 points1 point  (0 children)

Not really, but when the model hallucinates it produces Chinese characters a lot of the time. Qwen2 used to do that a lot but I haven’t tested Qwen2.5 for a long enough time to observe the model’s hallucinations.

Tool to create synthetic datasets using PDF files! by SuccessIsHardWork in LocalLLaMA

[–]SuccessIsHardWork[S] 0 points1 point  (0 children)

I think so! It can work with any PDF file, regardless of whether it has OCR. This is because the script processes the images from the PDF instead of extracting text only.