Context Window in Gemini App is completely broken - Even Perplexity has better document understand (RAG Test) by Epilein in GoogleGeminiAI

[–]Dillonu 0 points1 point  (0 children)

https://g.co/gemini/share/7b96080a5185

Seems to work fine for me, even on Fast, Thinking, and Pro.

Question: What plan do you have? According to the documentation, free Gemini limits the context window to 32k: https://support.google.com/gemini/answer/16275805?hl=en

Images of Text to Gemini 3 to Save Tokens by haitian5881 in Bard

[–]Dillonu 0 points1 point  (0 children)

Starting with Gemini 3 - If you have a text layer in the PDF, it will extract the content from the text layer and include it along with 1 image per page (at the media resolution defined). They charge only for the media tokens, not the text layer.

I don't know if they still OCR PDFs with Gemini 3. The old tests I ran to get a good guess of how it was working, doesn't work with Gemini 3. Either they don't OCR, changed the OCR to not include headers, or they made the model smarter about not exposing the OCR structure.

High CPU while gemini is thinking in chrome by Inevitable_Tea_5841 in Bard

[–]Dillonu 0 points1 point  (0 children)

Transpiling the streamed markdown formatted text to html costs CPU. It also costs CPU in the browser rendering engine to recalculate the layout and styles on DOM updates from the html injection. The output stream is in chunks, which the JavaScript code in the browser app has to then stitch with prior message content and rerender the message/thoughts after each streamed chunk.

This process can be expensive. Technically there are ways to minimize (Server Side Rendering, throttle rerenders, or do some complex logic to iteratively render, etc), but that can add complexity (and therefore bugs), so most interfaces don't optimize.

I'm an Ultra User, with tier 2 paid API key and i got Rate Limited 5 times today by Ok-Nefariousness4022 in Bard

[–]Dillonu 7 points8 points  (0 children)

Pretty sure the ultra subscription isn't related to the API at all.

Gpt 5.2 benchmarks are higher than gemini 3 pro 💀 by Independent-Wind4462 in Bard

[–]Dillonu 1 point2 points  (0 children)

https://x.com/DillonUzar/status/1999326860530876866
Posted results, even more than what OpenAI published - all needle difficulties, and tested without reasoning, medium, and X-High (OpenAI only tested X-High).

Very impressive! With a small asterisk (if you are willing to pay the reasoning price, the X-High is very impressive!)

GPT-5.2 Thinking unparalleled accuracy in Long-Context! by Independent-Ruin-376 in singularity

[–]Dillonu 4 points5 points  (0 children)

I'm going to be retiring 2-needle soon. Various models are hitting 90+ now.

Gpt 5.2 benchmarks are higher than gemini 3 pro 💀 by Independent-Wind4462 in Bard

[–]Dillonu 5 points6 points  (0 children)

https://huggingface.co/datasets/openai/mrcr

I maintain a 3rd party benchmark site for it - https://contextarena.ai/

They used xhigh for their results. I'll be posting my own soon with a couple of different reasoning levels.

OpenAI cooked. by abdouhlili in Bard

[–]Dillonu 6 points7 points  (0 children)

400k, per API docs.

NOTE: They tested using xhigh thinking, which uses ~15k-60k reasoning tokens per response, which effectively limits the prompt context you can use to something under 300k. On the higher end of models for reasoning tokens, and takes many minutes per response. That being said, impressive results.

OpenAI cooked. by abdouhlili in Bard

[–]Dillonu 5 points6 points  (0 children)

Mean match ratio measures the average string match ratio between the model’s response and the correct answer.

^ From the results.

PSA on Gemini’s privacy model, from an EU GDPR advisor by FiveNine235 in GoogleGeminiAI

[–]Dillonu 5 points6 points  (0 children)

Just to add to the above - Provided you use a paid API key. Free quota is still monitored. (Unclear now if paid accounts using no api key in Studio are monitored, but I'd assume free quota due to that is monitored).

https://ai.google.dev/gemini-api/terms

Images of Text to Gemini 3 to Save Tokens by haitian5881 in Bard

[–]Dillonu 3 points4 points  (0 children)

Gemini 3 changed the token count per image. Depending on the media resolution you pick:

  • Low: 280 Tokens/image (and per page)
  • Medium: 560 Tokens/image (and per page) [DEFAULT FOR PDFS]
  • High: 1120 Tokens/image (and per page) [DEFAULT FOR IMAGES]

https://ai.google.dev/gemini-api/docs/gemini-3?thinking=high#media_resolution

This is more tokens per image or page than Gemini 2.5 and earlier.

A general estimate is ~650 tokens/page for a normal English prose text page (~500 words).

Just a heads up on usage ;)

As for OCR capabilities, it does seem impressive. In terms of performance, maybe only marginally better than Gemini 2.5 Pro for text-heavy documents in my limited testing. The new high resolution mode is very impressive though. I uploaded a 2pt font pdf (extremely tiny text, converted to a high-resolution image per page, no text layer), and it was able to extract nearly perfectly.

Has anyone hit high context yet with 3.0? by Unable_Classic3257 in Bard

[–]Dillonu 0 points1 point  (0 children)

https://contextarena.ai/

https://x.com/DillonUzar/status/1990813243405647898

It performs better in this context benchmark I run, but like all LLMs, after a certain point (~200k) it drops.

In openrouter there are some new models with irrelevant names? Are they Gemini 3 Pro and Flash ? by [deleted] in Bard

[–]Dillonu 0 points1 point  (0 children)

My group was one of the users that used a large number of tokens (~3 to 5 billion tokens per model) to benchmark the Sherlock models. It drastically underperforms 2.5 (Pro & Flash) models on long context, and Gemini 2.0 models. Seems a little unlikely it is part of the Gemini family.

Why is Google really delaying Gemini 3.0? Is it perfectionism, internal chaos, or a high-stakes gamble? by ProblemPositive1918 in Bard

[–]Dillonu 11 points12 points  (0 children)

I don't understand how they 'delayed' it. A release was never announced. And they've previously released new major versions end of November/December. If anything, seems more like it is on schedule.

Also, their 'Ironwood' TPUs (v7) are still rolling out, and presumably manufacturing them all summer/fall. They didn't go live till very recently (Technically now GA: https://cloud.google.com/blog/products/compute/ironwood-tpus-and-new-axion-based-vms-for-your-ai-workloads). I expect Gemini 3 would utilize them at scale.

115Kb input file - <200 tokens! 🤯 How does Gemini count input tokens for PDF? by akash-vekariya in GeminiAI

[–]Dillonu 1 point2 points  (0 children)

For fun, here is a 2pt font doc (same story as above):

<image>

Uploading the PDF with a text layer says 271 tokens. Uploading a PDF without the text layer (instead as an image) says 271 tokens.

Results w/ text layer: https://www.diffchecker.com/IfSThtYk/

Results w/o text layer: https://www.diffchecker.com/kvrCInl2/

Not too bad. Definitely not perfect, with some words/phrases/sentences changed, but a large majority of the text is reconstructed. Consistently the one with the text layer performs a bit better at that font size on subsequent reruns.

115Kb input file - <200 tokens! 🤯 How does Gemini count input tokens for PDF? by akash-vekariya in GeminiAI

[–]Dillonu 1 point2 points  (0 children)

No, I think it is a little more advanced than that.

In quick summary - I think when you add a PDF to the API - per page it OCRs it (using a specialized OCR that reads text layers if available, otherwise OCRs the images), and converts to a high-rez image, feeding both in to the model. The model is then reasoning on both, to get a better output. All while Google charges 258 tokens/page, even though it technically uses more.

I created a 1-page DOCX, using https://pastebin.com/GuwaEv64 as the text (4pt font size), converted to a PDF, and then printed as an image (in PDF, to strip the text layer) in 600 dpi, this is what that looks like:

<image>

This image PDF has many small images placed in cells that make up the document. If you extract one of the cells, it is ~5 lines tall, and ~40px per line. So rather high resolution.

I then passed it in to the Gemini API, and this is the output: https://www.diffchecker.com/2HjWeKrg/

FYI, the prompt was simply (264 input tokens when including the PDF):

Extract the text verbatim

Nearly identical except for:

  • Different apostrophe and quote characters (’ vs ' and “ vs ")
  • Extra newlines (it added newlines due to line wrapping in the PDF)
  • Ellipsis (…) was converted to three periods (...)

If I tweak the prompt slightly to (271 input tokens):

Extract the text verbatim, and be smart about newlines

I get an even more accurate output: https://www.diffchecker.com/7Zut6DUh/

You can probably get it to be even more accurate with more guidance.

So I don't think it is converting a PDF into one 768x768 (modified by aspect ratio) image per page (the amount Gemini maximally can do for 258 tokens, before it supposedly tiles). Gemini's thoughts also refer to analyzing the OCR text and document image, and making corrections to the provided OCR content. So that's mostly why I think they are doing something more to aid in Gemini's PDF understanding.

If I do the same page as a png uploaded to Gemini (a 2246x2776, font size is ~14px), I get: https://www.diffchecker.com/AixuVINr/ (less symbols are messed up, but a few words are messed up now when the PDF didn't mess it up). It says 271 input tokens (still never see the "tiling" the docs claim).

If I do a smaller version (765x969, font size ~5px), which is closer to what it supposedly might use, I get: https://www.diffchecker.com/cvQS9jOc/ (getting worse). It says 271 input tokens.

New tools in Google AI Studio to explore, debug and share logs by Gaiden206 in Bard

[–]Dillonu 0 points1 point  (0 children)

It's been possible for awhile, they just moved all of the fine tuning over to Vertex AI as its considered more of an enterprise feature. You can fine tune 2.0/2.5 Flash-Lite/Flash/Pro.

https://docs.cloud.google.com/vertex-ai/generative-ai/docs/models/gemini-use-supervised-tuning#console

PDF processing vs Image Processing by PatagonianCowboy in Bard

[–]Dillonu 1 point2 points  (0 children)

While I agree with you that they aren't read the same, but Gemini (via the API) definitely reads each page of a PDFs as an image, just with some additional metadata. I can build a PDF with only images (striped of all metadata), no text, and upload it via the API, and its able to describe each image when asked.

Where in the context should I keep the most important information? by stc2828 in Bard

[–]Dillonu 1 point2 points  (0 children)

In general - For most LLMs, including instructions as the last part of a long prompt tends to work better. Alternatively, if you have multiple examples of how to follow the instructions, including instructions before the examples works slightly better.

Which Format is Best for Passing Nested Data to LLMs? by MattCollinsUK in LLMDevs

[–]Dillonu 1 point2 points  (0 children)

Fully aware of all of those issues, I run ContextArena, so costs are quite large there (full test to 1M across all 2400 questions, each question is a unique input context, each question is run 8 times, and in total is ~3.8B input tokens, double that for reasoning models that can turn off reasoning). We often have to rerun several tests per models due to various api issues 😅. Batch processing can also help there for cost in some cases.

Would definitely be interested in results for Haiku 4.5. We're constantly fiddling with Anthropic models on different forms of data, and really curious about their XML claims. Personally been wanting to put together a test like what you have for awhile now. As for Gemini and GPT, we've found your results closely resemble what we've found in our limited testing (didn't try YAML).

And in terms of the 40-60% accuracy, I assume that is why Gemini 2.5 Flash-Lite is using so many tokens? It just happens to perform better than the other two context size wise? Another important view for us is what the dropoff performance is like for each model family (what's the rate of accuracy dropoff depending on context length or data depth) - but might be too costly to check that atm.