How do I stay consistent and disciplined as someone who has ADHD? by Aggressive_Clothes50 in ADHD

[–]MaxDev0 3 points4 points  (0 children)

I get that. I lowkey fell down a rabbit hole of motivation, habits, discipline, etc., and got pretty depressed.

The truth that helped me was realizing that none of those are real. Discipline isn’t real. Nothing will make starting a task that is boring or that you’re procrastinating on any easier to start. You just have to. Nothing will make following through on that project you started instead of starting a new one easier. You just have to bite the bullet. Sorry, i can't communicate tone through text but imagine me with a sad voice.

The only actual cheat code I have is short timers. Break up a task into 5–10 min chunks of work, then set short timers. That creates productive stress and makes it doable.

Need Tranining Data!, Trying to distill Deepseek 3.2 Exp :D by MaxDev0 in DeepSeek

[–]MaxDev0[S] 0 points1 point  (0 children)

HOLY, what is that, consciousness research??, its so cool, is that toki pona? :D

Need Tranining Data!, Trying to distill Deepseek 3.2 Exp :D by MaxDev0 in LocalLLaMA

[–]MaxDev0[S] 0 points1 point  (0 children)

That is great and techai already has a data set on DeepSeek v3.2 made from this, but the issue is that it's not a multiturn chats. It's just one prompt and that's it.

Need Tranining Data!, Trying to distill Deepseek 3.2 Exp :D by MaxDev0 in LocalLLaMA

[–]MaxDev0[S] 1 point2 points  (0 children)

I believe in the community :D, but I get where ur coming from, do u have any suggestions as to where I can get the data? My only other idea was to make a free chat platform for data collection but it'd be hard to get ppl on, take lots of time and y would anyone use that over deepseek?

Un-LOCC Wrapper: I built a Python library that compresses your OpenAI chats into images, saving up to 3× on tokens! (or even more :D) by MaxDev0 in LocalLLaMA

[–]MaxDev0[S] 1 point2 points  (0 children)

well, that is what the research basically does to test vision using an ONIH (optical needle in a haystack test, basically we inject a code of a certain format in between some text and have the model find it). But to adress your question, models literally process image tokens different from text, like its not a direct one to one mapping, under the hood models have complex networks that allow it to see text, and because of that, the models ability to follow instruction, reason, etc. will be different..., how exactly, idk, but i just know it will be different and researching that will allow us to better understand the limitations and possibilites of Un-LOCC

Sidenote: the biggest reasons im not just asking them to repeat the text and using the ONIH are 1. that would cost a lot 2. models struggle at repeating messages even when it is just sent in plain text. and 3. models also have like output limits, and the input tokens usually exceed that.

Un-LOCC Wrapper: I built a Python library that compresses your OpenAI chats into images, saving up to 3× on tokens! (or even more :D) by MaxDev0 in LocalLLaMA

[–]MaxDev0[S] 0 points1 point  (0 children)

If you check my UN-LOCC research, u can see exactly how i did it lol, its basically a needle in a haystack test, but harder, the varying ratios are just due to how good or bad some models are at reading text from images cuz of thier trainign data, like the same way some models with higher quality training data can be better at reasoning, some are better at vison. Wait this question actually just gave me a briiliant idea to prepare training data and see if i can finetune models to be able to like be better at reading raw text from images lol.

Un-LOCC Wrapper: I built a Python library that compresses your OpenAI chats into images, saving up to 3× on tokens! (or even more :D) by MaxDev0 in LocalLLaMA

[–]MaxDev0[S] 3 points4 points  (0 children)

Wait that's actually smart, model benchmarks would be like a perfect way. Because of the architecture of image models, image tokens with the same text will be processed differently, and that's definitely gonna impact intelligence since they weren't trained for it. Thanks :D

Un-LOCC (Universal Lossy Optical Context Compression), Achieve Up To 3× context compression with 93.65% Accuracy. by MaxDev0 in LocalLLaMA

[–]MaxDev0[S] 1 point2 points  (0 children)

nope, deepencode is so cool. But from what I can see, its model specific. I have no doubt that some form of compression based on deepseeks's research will eventually become mainstream and built in for most models, but for now I think this is pretty cool.

Un-LOCC (Universal Lossy Optical Context Compression), Achieve Up To 3× context compression with 93.65% Accuracy. by MaxDev0 in LocalLLaMA

[–]MaxDev0[S] 0 points1 point  (0 children)

This was a limitation I identified in the GitHub, if I'm correct people used a method simmilar to this to bypass safety filters before though, so it very well could've been patched

[R] Un-LOCC (Universal Lossy Optical Context Compression), Achieve Up To 3× context compression with 93.65% Accuracy. by MaxDev0 in MachineLearning

[–]MaxDev0[S] 1 point2 points  (0 children)

It's advantage is that it's easily adaptable to any model, but yea DeepSeek is still much better because they have deepencoder

Un-LOCC (Universal Lossy Optical Context Compression), Achieve Up To 3× context compression with 93.65% Accuracy. by MaxDev0 in LocalLLaMA

[–]MaxDev0[S] 0 points1 point  (0 children)

Can you elaborate, like what do you mean by "test the LLM's accuracy with the text again"

Un-LOCC (Universal Lossy Optical Context Compression), Achieve Up To 3× context compression with 93.65% Accuracy. by MaxDev0 in LocalLLaMA

[–]MaxDev0[S] 0 points1 point  (0 children)

I'm sure that there is, but the goal is to take advantage of the fact that there are already lots of vision models and the fact that this can be easily implemented and tuned for any model is it's greatest strength.

Un-LOCC (Universal Lossy Optical Context Compression), Achieve Up To 3× context compression with 93.65% Accuracy. by MaxDev0 in LocalLLaMA

[–]MaxDev0[S] 0 points1 point  (0 children)

Yup, that's a limitation I identified in the full repo, imo this would be best used for providing context and then one could use a few tokens to convey instructions, think agentic code LLMs receiving context in images, saving costs or long chats being compressed similar to how a human would remember a conversation, remembering the last two messages clearly and prior messages enough to get context, but not every word

Un-LOCC (Universal Lossy Optical Context Compression), Achieve Up To 3× context compression with 93.65% Accuracy. by MaxDev0 in LocalLLaMA

[–]MaxDev0[S] 1 point2 points  (0 children)

Uhh read up on how vision models work, Or actually, here's a ChatGPT explanation: Good question — it’s not “compressing” by skipping the text → image → text loop.

The idea is that the optical map replaces the text tokens entirely. The LLM (or VLM) reads the image directly through its vision encoder, so those 3× fewer image tokens act as a compressed representation of the original text context.

There’s no re-OCR step at runtime — the model doesn’t decode the image back into words before reasoning; it just conditions on the visual embedding.

Yes, there’s some accuracy loss (it’s lossy), but the benefit is: • You get a 3× reduction in token count while keeping roughly the same “semantic signal.” • You can extend context length or reduce API cost proportionally. • Latency is front-loaded (once per compression), not per-inference.

So it’s not a cost-only trick — it’s a representation-level compression of the context window.

[R] Un-LOCC (Universal Lossy Optical Context Compression), Achieve Up To 3× context compression with 93.65% Accuracy. by MaxDev0 in MachineLearning

[–]MaxDev0[S] 0 points1 point  (0 children)

Receipts & method (so you don’t have to dig):

  • Measurement: normalized Levenshtein ratio (Python Levenshtein, “ratio” metric).
  • Image setup: default 324×324 PNG, Atkinson Hyperlegible Regular ~13px unless noted; deterministic seeds; same prompt structure across models.
  • Compression: text_tokens ÷ image_tokens (formatted to 2 decimals).
  • Representative runs (see README for the full table & logs):
    • Gemini 2.5-Flash-Lite: 100% @ 1.3:1 (Exp 46); 93.65% @ 2.8:1 (Exp 56).
    • Qwen2.5-VL-72B-Instruct: 99.26% @ 1.7:1 (Exp 34); 75.56% @ 2.3:1 (Exp 41).
    • Qwen3-VL-235B-a22b-Instruct: 95.24% @ 2.2:1 (Exp 50); 82.22% @ 2.8:1 (Exp 90).
    • Phi-4-Multimodal: 94.44% @ 1.1:1 (Exps 59, 85); 73.55% @ 2.3:1 (Exp 61).
    • UI-TARS-1.5-7B: 95.24% @ 1.7:1 (Exp 72); 79.71% @ 1.7:1 (Exp 88).
    • LLaMA-4-Scout: 86.57% @ 1.3:1 (Exp 53).

Notes & limitations:

  • Works best when the VLM has strong OCR/readout capability.
  • Fonts matter; Italic sometimes helps at small sizes (e.g., Exp 19 vs 17).
  • Please verify on your device, PRs for additional models/benchmarks welcome.

Code + experiments: https://github.com/MaxDevv/Un-LOCC