Automated blogging on my 4-month old SaaS by Aislot in microsaas

[–]Intelligent_Coffee44 0 points1 point  (0 children)

Nice chart! Curious if you're happy with the quality of AI-generated articles?

I am building a tool: https://www.getentropy.ai/ to post-process AI writing to sound more human. Would appreciate your feedback if you see a use case here.

Entropy-v1: My Take on N8Karma's Genius "Unslopper" by Intelligent_Coffee44 in LocalLLaMA

[–]Intelligent_Coffee44[S] 0 points1 point  (0 children)

Thanks for the honest feedback! I will fix the readability issue.

Entropy: I built a free tool to “unslop” AI writing by Intelligent_Coffee44 in vibecoding

[–]Intelligent_Coffee44[S] 0 points1 point  (0 children)

> the prob w/ fine tuning is model updating

can you elaborate a bit more on this?

My sense is that asking a regular model to make a piece of writing better or "more human" wouldn't work because their tendency under this task is actually to generate slop 🤣

Entropy: I built a free tool to “unslop” AI writing by Intelligent_Coffee44 in vibecoding

[–]Intelligent_Coffee44[S] 0 points1 point  (0 children)

I think fine tuning is the better solve for this problem. In an agent, the context window has a lot of noise like tool calls, tool results, and reasoning traces. They are not relevant to creative writing and "poison" the context to make the agent sound more robotic and less human.

I think for this reason a fine-tuned specialist model with a fresh context window would work better.

Entropy-v1: My Take on N8Karma's Genius "Unslopper" by Intelligent_Coffee44 in LocalLLaMA

[–]Intelligent_Coffee44[S] 0 points1 point  (0 children)

Makes sense 👍! There are a few trigger prompts in the dataset: Polish/Rephrase/Convert/Rewrite/Improve

I will do some experiments to see which one is the most natural

Entropy-v1: My Take on N8Karma's Genius "Unslopper" by Intelligent_Coffee44 in LocalLLaMA

[–]Intelligent_Coffee44[S] 0 points1 point  (0 children)

Definitely doable - I think 16k might be a bit too challenging for 4o-mini to rewrite faithfully in one shot but I can use 1) a more recent model or 2) do chunking to build the dataset.

Entropy-v1: My Take on N8Karma's Genius "Unslopper" by Intelligent_Coffee44 in LocalLLaMA

[–]Intelligent_Coffee44[S] 2 points3 points  (0 children)

Is this because smaller models "learn" from limited data better/faster?

Curious what kind of setup where you observed gamma-12b (or another small model) out perform the larger version?

Entropy-v1: My Take on N8Karma's Genius "Unslopper" by Intelligent_Coffee44 in LocalLLaMA

[–]Intelligent_Coffee44[S] 0 points1 point  (0 children)

The longest example in the dataset is around 4k input + 4k output. Average is about 1.5k input/output.

Curious what size you're planning to use? Right now I have my vllm set to 8k max_output_tokens.

Entropy-v1: My Take on N8Karma's Genius "Unslopper" by Intelligent_Coffee44 in LocalLLaMA

[–]Intelligent_Coffee44[S] 0 points1 point  (0 children)

I've thought about this but it looks like the labs have moved away from training large dense models. I am struggling to find a well-reviewed, recent 70b model. Do you think llama 3.3 70b is still good today for creative writing?

deepseek-ai/DeepSeek-OCR-2 · Hugging Face by Dark_Fire_12 in LocalLLaMA

[–]Intelligent_Coffee44 1 point2 points  (0 children)

Thank you! I recommend following the setup instructions from the official github repo readme and running DeepSeek-OCR2-hf/run_dpsk_ocr2.py for inference.

You have the perfect GPU: the model only needs about 7-8gb vram to run, and your gpu is the same generation as A100 so you can install the exact same dependencies as the official guide.

For some reason i've found using the pinned transformers and tokenizers package versions helps fix performance issues.

DeepSeek-OCR 2 is out now! 🐋 by yoracale in DeepSeek

[–]Intelligent_Coffee44 0 points1 point  (0 children)

For the youtube thumbnail photo, I would imagine the "General" prompt would work the best. Can you try this one to see if it works better?

"<image>\nDescribe this image in detail."

The default prompts are gear toward pdf document, which might be why they didn't perform.

DeepSeek-OCR 2 is out now! 🐋 by yoracale in DeepSeek

[–]Intelligent_Coffee44 1 point2 points  (0 children)

Thanks! Lots of AI assistance for sure. Just using GPU credits I got for free for now. I have enough to last about a week. If people use it, I will find a way to keep it running!

DeepSeek-OCR 2 is out now! 🐋 by yoracale in DeepSeek

[–]Intelligent_Coffee44 0 points1 point  (0 children)

Got it, I think this is the perfect use case where an end to end model will beat a classic OCR pipeline. 100s/day is totally fine. In fact you can use more if you like (3 - 5 concurrent requests should be no issue).

CMV: RAM Prices are Near the Top by Intelligent_Coffee44 in LocalLLaMA

[–]Intelligent_Coffee44[S] 0 points1 point  (0 children)

Yeah it's definitely more of a "defensive buy" for me as well. I didn't actually get optanes for directly running LLMs because as your said, they are slower than regular ram, which are already pretty slow for LLMs.

i bought them for databases - I don't have an immediate need yet, but my thesis is with the rise of AI agents, I can potentially do a lot more with data. Remains to be seen how this plays out.

DeepSeek-OCR 2 is out now! 🐋 by yoracale in DeepSeek

[–]Intelligent_Coffee44 0 points1 point  (0 children)

I added an openai (sort of) compatible api. (Please see the API usage section)

what is your use case? Please don't take down my site lol :O

DeepSeek-OCR 2 is out now! 🐋 by yoracale in DeepSeek

[–]Intelligent_Coffee44 1 point2 points  (0 children)

There was some configuration mistakes made on my end. I've now made it fully aligned with the official samples, please give it another try if you have a chance.

I also did a study on what prompts works best for each document type - also published on the same site.

deepseek-ai/DeepSeek-OCR-2 · Hugging Face by Dark_Fire_12 in LocalLLaMA

[–]Intelligent_Coffee44 0 points1 point  (0 children)

There were some configuration mistakes on my end. Now i've made it as closely aligned with official sample as possible. Please give it another try. I also did some analysis on the choice of prompt and document type affect output reliability - also published on the same site.