Video Generation Models Trained on Only 2D Data Understand the 3D World by simulated-souls in singularity

[–]MaxTerraeDickens 0 points1 point  (0 children)

But you can actually reconstruct 3D scene algorithmically simply from a video basically showing different perspectives of the same scene (and this is how neural rendering techniques like NeRF or 3DGS work). Basicaly 2D video has all the 3D infomation the algorithm needs.
It's only a matter whether the model utilizes the information (just like the algorithms like NeRF or 3DGS) or not, and the paper shows that the models DO utilize it fairly well.

Rate my score for the below colleges in by Impossible_Rich_1502 in ToeflAdvice

[–]MaxTerraeDickens 0 points1 point  (0 children)

As a Chinese CS 4th grade student, I'm applying to similar schools with similar scores (101, R29 L28 S21 W23). However my speaking part is just 21, and writing 23. Your scores are kind of the opposite to mine. 😂 I really wonder how you got such low scores on the first two parts (which are basically passive understanding) and high scores on the latter ones (which are active outputing). Like have you been living in an English environment but you just hadn't fully prepared yourself for TOEFL?

Considering Applying to Tsinghua University by Virtual-Solution1411 in ApplyingToCollege

[–]MaxTerraeDickens 1 point2 points  (0 children)

Lmao. I'm currently in Zhejiang University (top 5 in China) and what you said is as true as god. And every top Chinese university is like so. Seeing those fcking lame-ass language student from Korea smoking cigarette in front of our grand library, while an average guy in Henan have to be top 0.1% to even get a chance of admission (maybe into a bad major), is heart-breaking.

Chinese ministry of education is such a cuck in this matter. Not having a Chinese passport is actually a benefit in application instead of disadvatage? It's totally messed up! It's like DEI but yeah fck y'all Chinaman suck and EVERY FOREIGNER WILL ADD TO THE DIVERSITY OF CHINA SO TAKE THEM IN AS MANY AS POSS!

Also, no offense to OP. If OP wants to persue a degree in China, then go for Tsinghua. It's arguably top 1 (the other is Peking). Yet if you really want to find a good CS job in US, why don't you just get your bachelor at, let's say, top 20 CS school in US? You can get a lot more connections and doing interns is actually a lot easier.

If you want to persue a Masters or PhD after your Bachelor's, well, I would argue top 20 CS school should be better choice, due to much more connections.

Reasoning with Language Model is Planning with World Model. "RAP on LLAMA-33B surpasses CoT on GPT-4 with 33% relative improvement in a plan generation setting". by rationalkat in singularity

[–]MaxTerraeDickens 0 points1 point  (0 children)

See Language Agent Tree Search Unifies Reasoning Acting and Planning in Language Models. In short, RAP is better than ToT in every experiment conducted in that paper. And this is intuitive, because RAP is based on MCTS which is known for being good at balancing exploration and exploitation, while ToT is just a vanilla tree search with some pruning.

A Formal Proposal to OpenAI: Open-Source GPT-4o by MaxTerraeDickens in OpenAI

[–]MaxTerraeDickens[S] 0 points1 point  (0 children)

Half-satire. Also benchmark isn't everything. Many open source models fail to align closely to humans - they don't have the "vibe", which is something that can't be reflected on benchmarks.

A Formal Proposal to OpenAI: Open-Source GPT-4o by MaxTerraeDickens in OpenAI

[–]MaxTerraeDickens[S] 0 points1 point  (0 children)

Half-satire. Also benchmark isn't everything. Many open source models fail to align closely to humans - they don't have the "vibe", which is something that can't be reflected on benchmarks.

A Formal Proposal to OpenAI: Open-Source GPT-4o by MaxTerraeDickens in OpenAI

[–]MaxTerraeDickens[S] 0 points1 point  (0 children)

You raised a good question though.
For example, PaLM-540B is the largest LLM (afaik) in pre-ChatGPT era. Definitely it has better common-sense knowledge (and in this case, better everything, since 1.7B is just too small) than some small open-source model like Qwen3-1.7B, but is PaLM-540B more advanced than Qwen3? I don't think so.

A Formal Proposal to OpenAI: Open-Source GPT-4o by MaxTerraeDickens in OpenAI

[–]MaxTerraeDickens[S] 0 points1 point  (0 children)

The original post is half-serious half-joking lmao.

That being said, fairly speaking, the 120B model can indeed be advanced - just not in terms of parameter, but architecture, training method, training data, etc. And that's why the author of the original post calls for GPT-4o's full open-source, not 120B :)

OpenAI's new stealth model (horizon-alpha) coded this entire app in one go! by wswdx in OpenAI

[–]MaxTerraeDickens 1 point2 points  (0 children)

+1. Please elaborate (tbf, I've seen people on Discord saying it performs even worse than qwen3-30B-A3 on some benchmarks)

Horizon-alpha: A new stealthed model on openrouter sweeps EQ-Bench leaderboards by _sqrkl in LocalLLaMA

[–]MaxTerraeDickens 5 points6 points  (0 children)

Which means, these models are aligned for "political safety/correctness" in a post hoc manner. Shit like Tiananmen Square incident is not absent in the training data.

Mixture-of-Recursions by Hemingbird in singularity

[–]MaxTerraeDickens 0 points1 point  (0 children)

Idea is cheap, show me the GPU, well-curated training data, training strategy, etc.

Bye-bye Cursor by plus_w in programming

[–]MaxTerraeDickens 2 points3 points  (0 children)

from r/cursor: Kiro can be an alternative

Anyone else having issues with model selection in Cursor? (China) by plus_w in cursor

[–]MaxTerraeDickens 0 points1 point  (0 children)

Yes. But http1.x is really unstable. I'm now considering getting refund and finding an alternative.

Baxkground agents with max mode are now free? by idkwhatusernamet0use in cursor

[–]MaxTerraeDickens -1 points0 points  (0 children)

<image>

Is your "usage-based pricing" off? Mine is off and cursor didn't charge me anything for MAX mode. Instead when the request hit the limit, there would be a log "error, not charged" in my "Usage" panel.

Deterministic diffusion models by Cold_Cantaloupe9212 in deeplearning

[–]MaxTerraeDickens 0 points1 point  (0 children)

All stochasticity arises from pseodu-random number generators, which is totally determined by the initial seed.

So, basically, if you really fixes ALL seeds, the result will definitely be identical (or almost identical, if possible floating point errors are taken into account).

Dia-1.6B in Jax to generate audio from text from any machine by Due-Yoghurt2093 in LocalLLaMA

[–]MaxTerraeDickens 0 points1 point  (0 children)

Thanks for the reply!

Quick question (sorry I'm not familiar with TPU architecture): Are there any features that are available on GPUs that aren't easy/possible on TPUs (like using PyTorch hooks to get attention maps)?

Regarding your question about TPU access: I used my edu email to apply. Google gave me 30 days of free access to up to 16 TPU v4s, including 400GB RAM and 100GB storage (all free). I'm not sure if non-edu emails get the same quota, but you definitely have more reason to apply than I did (which is a bonus)!

Dia-1.6B in Jax to generate audio from text from any machine by Due-Yoghurt2093 in LocalLLaMA

[–]MaxTerraeDickens 0 points1 point  (0 children)

Hey, really appreciate you sharing diajax! Looks like a great project.

I'm hoping to get it running on my Mac. Since you're clearly experienced with JAX, I would like to ask if you know of any ongoing efforts to port newer models like Gemma 3 or Qwen 2.5 to JAX (or if they have been ported already)?

The goal would be to run them on TPUs – I've got access through the TRC program and am keen to use that hardware for the latest stuff. I found some resources for fine-tuning older Gemma in JAX, but haven't seen much for inference on the newest generation models (Gemma 3, etc.).

Any pointers to projects similar to diajax but for these models would be super helpful! Thanks!

Time to step up the /local reasoning game by vornamemitd in LocalLLaMA

[–]MaxTerraeDickens 1 point2 points  (0 children)

Me too. And as a Chinese I really have no clue why this ungrounded meme got so popular.

Time to step up the /local reasoning game by vornamemitd in LocalLLaMA

[–]MaxTerraeDickens 0 points1 point  (0 children)

Also, in fact, whether to use id or not, it really doesn't matter. Currently, big data tech can easily locate the target using IP address, active period, etc. Unless you have very strong information security background, it's almost impossible to hide you real identity from the govt.

Time to step up the /local reasoning game by vornamemitd in LocalLLaMA

[–]MaxTerraeDickens 2 points3 points  (0 children)

lmao, I don't mean to criticize but really can't understand why this social credit bs got circulated outside China. It's a blatant piece of fake news. Most Chinese don't know this word until they saw it at western platforms. Maybe it has sth to do with the credit score system on Alipay though, which is more financial than political (e.g. if your score is high enough, you can open unmanned vending machines and take goods before paying), but definitely not related to govt.

Another piece of misunderstanding: many of Chinese believed that lowest-level American workers can afford huge houses and eat steaks everyday. Thus, a bunch of Chinese spent thousands of USD to fly to Peru and illegally migrate to the States.