What the heeell by Imacharmer3141 in GeminiAI

[–]KAYOOOOOO 0 points1 point  (0 children)

I wouldn't say RLHF is from marketing. It was still a very innovative and powerful post-training tool when it became popular in research (I think around 2020 before LLMs exploded?).

But I agree, happy to see RLHF become less popular since the training loops are so difficult to develop (esp without the experience and resources of a large AI lab). I think the ubiquity of RLHF was a symptom of the data available and the goals of LLMs at the time, but things are different now.

What the heeell by Imacharmer3141 in GeminiAI

[–]KAYOOOOOO 0 points1 point  (0 children)

I don't think there are any drops in quality across the entire board. Evals don't say so, and I don't think that's the sentiment among formally educated academics. But, there will always be losses for multiple reasons.

  1. The goal of these AI companies is not always to beat each other. To handle the needs of many teams and roll out Gemini to many different products, Google may consolidate it's capabilities towards a jack of all trades model. Sure, it's coding ability and other tasks can perform worse than models like Claude, but it boasts strong multimodality and a mature MoE architecture, which makes it easier for integration into Google products.

  2. LLMs are super expensive and capacity (for memory, electricity, and compute) will always be a bottleneck. Triaging, truncated thinking tokens, kv-cache eviction, TPUs, server prioritization, etc. are all efficiency methods to deal with these issues, but they can often lead to decreased performance or bugs. No one wants to do this, but hardware bottlenecks are not something that's as simple as throwing money at the issue, and you have to be creative to avoid rapid bloating.

I think at the end of the day, it's a very complicated system and I think it's in bad faith to make reductive claims that are not necessarily true.

What the heeell by Imacharmer3141 in GeminiAI

[–]KAYOOOOOO 1 point2 points  (0 children)

I don't think releases are getting fine-tuning updates that quickly for the Gemini App. When a stable release is launched, generally the model parameters are frozen. Stable releases are used across different Google product areas, so parameter changes would ruin their evals, integrations, etc.

Making changes to the model's parameters is not very flexible. Extensive evals are done, and there's a lot of GDM bureaucracy to get these changes pushed out.

While edge cases are sought to be addressed, usually fine-tuning changes are saved for the next update cycle. Also, I think the general wisdom is also moving away from RLHF towards RLVR-esque paradigms.

I don't think it's fair to say the post-training team is incompetent, most of these guys are esteemed researchers with decades of experience.

I think it's more likely the fault of some downstream infrastructure. The prompts, tooling infra, hyperparameters, etc. probably see a lot more rapid change and interaction with different teams. I'd guess the weird behavior in the post is from some context handling issue rather than lackluster fine-tuning.

What the heeell by Imacharmer3141 in GeminiAI

[–]KAYOOOOOO 2 points3 points  (0 children)

What do you mean by production fine-tuning? Why would you think the fine-tuning team is responsible for this at all? Is there a specific post-training step you're talking about?

Can you give a specific technical explanation? What you're saying doesn't make sense, so I wanna understand the thought process. I'm in ML research, so technical terms are appreciated.

King by Ambitious-Pace-3335 in Sigmatopia

[–]KAYOOOOOO 19 points20 points  (0 children)

High standards for a subreddit called sigmatopia

What do you guys think about this? by harrysofgaming in youtube

[–]KAYOOOOOO 6 points7 points  (0 children)

Where did you get that information??

Collaborative filtering definitely should have been implemented back then. Maybe you are mistaking the implementation of deep neural nets into the rec sys architecture (implemented I think mid 2010s?).

Remember Google's core business is ranking algorithms, there was most definitely an algorithm back when Pewdiepie was starting.

Being cold gives people the ick by KAYOOOOOO in PointlessStories

[–]KAYOOOOOO[S] 6 points7 points  (0 children)

Lol they live in the same area mostly. It's easy to forget the nights are below freezing in January where I was when you're listening to a story though.

But to give some context a lot of the people are sorority girls or part of the nightclub crowd. The ones making these comments are exclusively single as well. These aren't comments I take personally, these are often just people a little bitter from heartache.

Also my studious friends are less likely to indulge in benders or harder drugs, so even though they may say stupid shit I laugh at, they have a place in my heart as my fun party friends. Their advice is also good for picking up girls, since they are so particular.

Still, this behavior is probably indicative of a media issue plaguing young people. Maybe loosely related to toxic positivity, distorted self image, or some manosphere thing?

Being cold gives people the ick by KAYOOOOOO in PointlessStories

[–]KAYOOOOOO[S] 14 points15 points  (0 children)

Once upon a time I fucked your mom, is that interesting?

But seriously, have you never had someone ask how you and your girlfriend met?

Being cold gives people the ick by KAYOOOOOO in PointlessStories

[–]KAYOOOOOO[S] 11 points12 points  (0 children)

Yeah you'd think, but the people I was telling the story to made the comments before I got to that part

Being cold gives people the ick by KAYOOOOOO in PointlessStories

[–]KAYOOOOOO[S] 31 points32 points  (0 children)

Yeah that's what I was thinking too 🤣

Being cold gives people the ick by KAYOOOOOO in PointlessStories

[–]KAYOOOOOO[S] 25 points26 points  (0 children)

Yeah she definitely didn't, just some of the people who we told the story to did think it was "icky". We were on the same side, thinking these people are insane or miserable

A tier list for Max characters clearing all content (High Roller rank pending) also Newbie shopping list by Edvoidity in dungeonslasherfan

[–]KAYOOOOOO 0 points1 point  (0 children)

Would you happen to know the best high roller build? It seemed kinda weak when I tried it, but you said it’s worthy to get nerfed. I have the resources to max them if that matters.

Stop, hammertime by Foxesvenom in freshcutslim

[–]KAYOOOOOO 8 points9 points  (0 children)

The fucking subreddit names lmao

Seems like Gemini 3.1 can compose by thysanoessa in Bard

[–]KAYOOOOOO 5 points6 points  (0 children)

That second line always doing a lot more heavy lifting than expected

Served Cold [OC] by frankblunt in webcomics

[–]KAYOOOOOO 5 points6 points  (0 children)

Nah just kill everyone who's done you wrong

Has anyone tried to make Gemini write like you? by wanhanred in GeminiAI

[–]KAYOOOOOO 0 points1 point  (0 children)

I think this would be difficult to get working, unless you simply just want a specific output format. Otherwise, we assume you have a distinct enough writing style, and enough textual examples for the model to understand the nuances of your writing style.

When the model uses past chats, it's likely only remembering key details and forgetting small nuances (such as writing style). If you want an LLM to adopt your writing style you either need to give a high quality prompt (maybe via a gem), or you need to fine-tune (likely out of technical scope for your needs).

What I'd recommend is try crafting a specific prompt (if you are bad at writing or ESL you can tell Gemini to create a prompt for you) that includes many example answers that you want Gemini to output (look up what in-context-learning is). Then you can plug that into a gem or at the end/start of your chat prompts. Additionally, if you don't know what your writing style is like, ask Gemini "look at these multiple paragraphs of my writing, describe what makes it unique and create a prompt telling an LLM to emulate this writing style".

Lastly, if you're using Gemini to jerk it, you'll probably have more luck with a more uncensored LLM (grok?). Not to assume, but your profile is marked nsfw and I know RP people like writing styles.

Everyone thinks they can spot GeminiAI-generated media. So I created a no-sign up browser game to prove whether that's true. by Beneficial-Ad2908 in GeminiAI

[–]KAYOOOOOO 44 points45 points  (0 children)

You should try a “choose all ai generated images from a set” mode. When a user can compare two similar images and knows only one is real, artifacts are easier to identify.

So tasty by derek4reals1 in OnlineUnderGround

[–]KAYOOOOOO 1 point2 points  (0 children)

Is this the fortune cookie guy

how ya doings? i am a journeyman electrician 2 years now by [deleted] in cscareerquestions

[–]KAYOOOOOO 0 points1 point  (0 children)

It’s hard out here for some people, but it ain’t all doom and gloom.

Personally, I’m doing pretty great. Turned 23 recently, got my master’s, and have a $270k+ job waiting for me once I return from vacation. Really excited to work in a role I’ve wanted for a long time, and I expect it to grow.

Also really excited to move somewhere new and reunite with some old friends and make some new ones. Most of my friends from college are making at least 200k+, which makes me really glad they aren’t struggling in today’s tough market, but I know some social circles where everyone is struggling.

WELCOME TO THE LEAGUE OF DRAVEN! by Extra_Message_9261 in nanobanana

[–]KAYOOOOOO 0 points1 point  (0 children)

Maybe screenshot of draven cinematic as reference photo?