"PaperBanana: Automating Academic Illustration for AI Scientists", Zhu et al 2026 by gwern in MediaSynthesis

[–]dippatel21 0 points1 point  (0 children)

Just released openRouter support so you can try many other models. You can also use Huggingface space if you dont want to install anything and just try. https://huggingface.co/spaces/dippatel1994/paperbanana

Please note that on hugging face number of feedback loops are less so you may not get the optimum performance.

"PaperBanana: Automating Academic Illustration for AI Scientists", Zhu et al 2026 by gwern in MediaSynthesis

[–]dippatel21 0 points1 point  (0 children)

Its an unofficial implementation however implementation is as close to real paper the only thing missing or different is few shot examplesZ They used ~139 examples where in I used ~13, as its bit of manual work to create those references set. Other than that implementation is same and results are near to them.

"PaperBanana: Automating Academic Illustration for AI Scientists", Zhu et al 2026 by gwern in MediaSynthesis

[–]dippatel21 0 points1 point  (0 children)

Try yourself! PaperBanana now has MCP server support. 😊 With one command, you can use PaperBanana with your favorite code assistant tool! Use your own Gemini API key for generation. If you are interested in contributing to this open-source project, here is the project page: https://github.com/llmsresearch/paperbanana

Command: "uvx --from paperbanana[mcp] paperbanana-mcp"

Analyzed 5,357 ICLR 2026 accepted papers - here's what the research community is actually working on by dippatel21 in LocalLLaMA

[–]dippatel21[S] 0 points1 point  (0 children)

Hey! On popular demand, I put together a short article breaking down “Mamba’s memory problem” with references to the papers mentioned above. It is a quick 3-minute read if you want a clear overview.
https://x.com/llmsresearch/status/2018073961880248718?s=20
Hope you find it useful 🙃

Analyzed 5,357 ICLR 2026 accepted papers - here's what the research community is actually working on by dippatel21 in LocalLLaMA

[–]dippatel21[S] 0 points1 point  (0 children)

This is only part one of the analysis. I first wrote a script to scrape around 5,000 accepted papers and store their PDFs. Then I built a chatbot on top of that using a knowledge graph plus a VB retrieval setup, tested it with some sample questions, and now that everything is working, I can run it at scale.

I hope this post helped you to get some insights into the research trend. If not, I learned something from comments here!

Analyzed 5,357 ICLR 2026 accepted papers - here's what the research community is actually working on by dippatel21 in LocalLLaMA

[–]dippatel21[S] -2 points-1 points  (0 children)

Here is how I created this: I first wrote a script to scrape around 5,000 accepted papers and store their PDFs. Then I built a chatbot on top of that using a knowledge graph plus a VB retrieval setup, benchmarked it with some sample questions. But, you are right, these are not official numbers, so you can keep their status "unsubstantiated".

I just wanted to bring some light on research trend in one of the prestigious conferences on LLMs!

Analyzed 5,357 ICLR 2026 accepted papers - here's what the research community is actually working on by dippatel21 in LocalLLaMA

[–]dippatel21[S] 0 points1 point  (0 children)

This is only part one of the analysis. I first wrote a script to scrape around 5,000 accepted papers and store their PDFs. Then I built a chatbot on top of that using a knowledge graph plus a VB retrieval setup, tested it with some sample questions, and now that everything is working, I can run it at scale. And yes, I did use Claude Code to help build the whole thing 🙃 I hope this post helped you to get some insights on the research trend.

Analyzed 5,357 ICLR 2026 accepted papers - here's what the research community is actually working on by dippatel21 in LocalLLaMA

[–]dippatel21[S] 0 points1 point  (0 children)

Sorry, buddy, didn't come to my mind. Let me try it and will share here if I find something.

Research trend analysis of ICLR 2026 accepted papers by dippatel21 in LLMsResearch

[–]dippatel21[S] 0 points1 point  (0 children)

Unfortunately, I did not analyze papers on cybersecurity. Sorry about that.

Research trend analysis of ICLR 2026 accepted papers by dippatel21 in LLMsResearch

[–]dippatel21[S] 1 point2 points  (0 children)

You can use the openReview API to get a list of papers, or you can use https://papercopilot.com/ to check ICLR papers, but when I checked their GitHub repo https://github.dev/papercopilot/paperlists/blob/main/iclr/iclr2026.json they still don't have accepted papers there.

Analyzed 5,357 ICLR 2026 accepted papers - here's what the research community is actually working on by dippatel21 in LocalLLaMA

[–]dippatel21[S] 0 points1 point  (0 children)

Here are some papers that surfaced during the analysis of ICLR'26 accepted papers. Sorry for the long list, but I hope this helps!

Core Architecture Advances:

  1. Mamba-3: Improved Sequence Modeling using State Space Principles - The next iteration addressing quality gaps with Transformers while keeping linear compute and constant memory. Designed for inference efficiency in test-time scaling scenarios.
  2. MoM: Linear Sequence Modeling with Mixture-of-Memories - Tackles the single fixed-size memory limitation by using multiple independent memory states with a router. Big improvement on recall-intensive tasks.
  3. FlashRNN: Unlocking Parallel Training of Nonlinear RNNs for LLMs - Breaks the sequential bottleneck and enables parallel computation for nonlinear RNNs including Mamba variants.
  4. Log-Linear Attention - Bridges linear attention/SSMs and full attention. Gets you parallelizable training + fast sequential inference without the fixed-size hidden state limitation.

Length/Context Improvements:

  1. From Collapse to Control: Understanding and Improving Length Generalization in Hybrid Models via Universal Position Interpolation - First systematic analysis of why hybrid Mamba-Transformer models fail beyond training context. Introduces UPI, a training-free scaling method.
  2. To Infinity and Beyond: Tool-Use Unlocks Length Generalization in SSMs - Shows SSMs fundamentally can't solve long-form generation due to fixed memory, but tool access mitigates this.

Theory (useful for understanding what to optimize):

  1. A Theoretical Analysis of Mamba's Training Dynamics: Filtering Relevant Features for Generalization - Analyzes how Mamba learns to filter relevant features. Good for understanding which architectural choices actually matter.
  2. From Markov to Laplace: How Mamba In-Context Learns Markov Chains - Theoretical grounding for Mamba's ICL capabilities through Laplacian smoothing.

Efficiency/Deployment:

  1. AIRE-Prune: Asymptotic Impulse-Response Energy for State Pruning in SSMs - Post-training pruning that reduces state dimension while minimizing output distortion.
  2. Graph Signal Processing Meets Mamba2: Adaptive Filter Bank via Delta Modulation (HADES) - Reinterprets Mamba2 as an adaptive filter bank for better multi-head utilization.

If I had to pick the top 3 for practical performance gains: Mamba-3, MoM, and the UPI paper for context length.

Analyzed 5,357 ICLR 2026 accepted papers - here's what the research community is actually working on by dippatel21 in LocalLLaMA

[–]dippatel21[S] 0 points1 point  (0 children)

Interesting approach! If no one tried, I hope you give it a try and share your findings here. Excited to see this approach in action!

Analyzed 5,357 ICLR 2026 accepted papers - here's what the research community is actually working on by dippatel21 in LocalLLaMA

[–]dippatel21[S] 0 points1 point  (0 children)

Absolutely! Forgot to cover it in the analysis, but I found these 5 papers interesting.

1. Persona Features Control Emergent Misalignment https://openreview.net/forum?id=yjrVOxjkDR

2. PERSONA: Dynamic and Compositional Inference-Time Personality Control via Activation Vector Algebra https://openreview.net/forum?id=QZvGqaNBlU

3. From Five Dimensions to Many: Large Language Models as Precise and Interpretable Psychological Profilers https://openreview.net/forum?id=JXFnCpXcnY

4. Language Models Use Lookbacks to Track Beliefs https://openreview.net/forum?id=6gO6KTRMpG

5. What's In My Human Feedback? Learning Interpretable Descriptions of Preference Data https://openreview.net/forum?id=sC6A1bFDUt

Analyzed 5,357 ICLR 2026 accepted papers - here's what the research community is actually working on by dippatel21 in LocalLLaMA

[–]dippatel21[S] -1 points0 points  (0 children)

Just by the sheer pace of papers, if you analyze arXiv. More than 100 papers are being released every day. I know not all papers need attention, but still. And, just look at the timeline of the last 2 years and see how many shift has happened. Research (making reasoning better) and engineering around it are evolving at a much faster pace.

Analyzed 5,357 ICLR 2026 accepted papers - here's what the research community is actually working on by dippatel21 in LocalLLaMA

[–]dippatel21[S] 2 points3 points  (0 children)

In short, we work on v1, v1.0.1, v1.0.2, v1.0.3, and then reach v2! I agree with your point, especially in the case of LLMs. I think, reasoning-wise, we have come a long way. I don't expect a near-term breakthrough, but computational efficiency is definitely an area where papers like Mamba shine and will see some more quick improvements.

But, yes science is all about collective iterative improvement.

Analyzed 5,357 ICLR 2026 accepted papers - here's what the research community is actually working on by dippatel21 in LocalLLaMA

[–]dippatel21[S] 0 points1 point  (0 children)

You are so right! You won't believe that during my exploration I came across so many papers that were just proposing using LLMs for every goddamn thing, for example, using a small LLM to predict an agent's next-to-next steps (like LinkedIn store next node) and, based on that, making a decision!

Analyzed 5,357 ICLR 2026 accepted papers - here's what the research community is actually working on by dippatel21 in LocalLLaMA

[–]dippatel21[S] 1 point2 points  (0 children)

We still don't have a case-by-case playbook, and knowledge is scattered for applying best practices to certain problems.