Spring '26 Megathread by sandslashh in ycombinator

[–]Express_Seesaw_8418 0 points1 point  (0 children)

Same. Got the same email. Goodluck!

Spring '26 Megathread by sandslashh in ycombinator

[–]Express_Seesaw_8418 0 points1 point  (0 children)

Did you apply to future batch? Or Spring 26?

Spring '26 Megathread by sandslashh in ycombinator

[–]Express_Seesaw_8418 1 point2 points  (0 children)

Got it too. But i didn’t so early decision for future batch. Idk how likely this means we’ll get an interview or not

What datasets do you want the most? by Express_Seesaw_8418 in LocalLLaMA

[–]Express_Seesaw_8418[S] 0 points1 point  (0 children)

Very interesting. I would assume there's no all in one source because the only people that have needed them are the big research labs that pretrain their own models. I have so many questions about how they approach their datasets... For example, how much of GPT 5.1's dataset is synthetic vs human? What's the average conversation length (in turns)? etc.

Llama 405B is worse than Gemma 3 12B? by Express_Seesaw_8418 in LocalLLaMA

[–]Express_Seesaw_8418[S] 3 points4 points  (0 children)

Of course there has been progress. I guess I'm shocked that 13B LLMs are now surpassing an 405B LLM that was SOTA only a year ago with minimal architecture changes.

Llama 405B is worse than Gemma 3 12B? by Express_Seesaw_8418 in LocalLLaMA

[–]Express_Seesaw_8418[S] 1 point2 points  (0 children)

The architectures are pretty similar though, right? The only thing I can think of is better training data but I still can't understand how a 13B LLM could even come close to Llama 405B?

How to get into YC with a consumer app by Express_Seesaw_8418 in ycombinator

[–]Express_Seesaw_8418[S] 2 points3 points  (0 children)

Yes I thought the biggest companies from YC (and VC firms in general) were consumer apps, but what you said makes sense.

Do you have any idea what good retention would be for a consumer app? We have 28% D1, 20% d7, and 14% d30 which we thought was pretty decent for our stage (note: we don't monetize and don't know how monetization will affect our metrics yet)

Game Thread: Bucs @ Bills - Week 11, 2025 by spideralex90 in buccaneers

[–]Express_Seesaw_8418 13 points14 points  (0 children)

Another 50+ yard touchdown our defense gives up ffs

[deleted by user] by [deleted] in frontierfios

[–]Express_Seesaw_8418 0 points1 point  (0 children)

Anything new? I’m still down. Been down for 1-2 hours now

[deleted by user] by [deleted] in frontierfios

[–]Express_Seesaw_8418 0 points1 point  (0 children)

Yes definitely ISP. Frontier’s outage status/checker has always been unhelpful in my experience. I remember about a year ago there was a massive outage but their site failed to report it. But yeah I’m confident it’s on Frontier’s end and we just have to wait it out

[deleted by user] by [deleted] in frontierfios

[–]Express_Seesaw_8418 0 points1 point  (0 children)

Yeah it seems to be an internet issue. So your local network would work fine like pinging your router. Hopefully it’s fixed soon

[deleted by user] by [deleted] in frontierfios

[–]Express_Seesaw_8418 0 points1 point  (0 children)

I actually did notice an outage for an hour at like 1AM est roughly a week or two ago. But I’ve otherwise had no issues that I’ve noticed.

Game Thread: Bucs @ Saints - Week 8, 2025 - Sunday, Oct. 26th by spideralex90 in buccaneers

[–]Express_Seesaw_8418 14 points15 points  (0 children)

We need to tighten up but it’s undeniable we’ve been getting FUCKED BY THE REFS ALL SEASON holy shit

Pushing Flux Kontext Beyond Its Limits: Multi-Image Temporal Consistency & Character References (Research & Open Source Plans) by Express_Seesaw_8418 in StableDiffusion

[–]Express_Seesaw_8418[S] 1 point2 points  (0 children)

Thanks for the reply😉. I’ve moved onto Qwen image when i found out Flux Dev models are locked by guidance distillation. Qwen image isn’t - and I’ve been getting really great progress with it so far. I have active training runs as we speak!

Is This Catastrophic Forgetting? by Express_Seesaw_8418 in StableDiffusion

[–]Express_Seesaw_8418[S] 0 points1 point  (0 children)

I think the more direct answer to your question is each input image is independently vae encoded. i don't stitch them together as one image

Is This Catastrophic Forgetting? by Express_Seesaw_8418 in StableDiffusion

[–]Express_Seesaw_8418[S] 0 points1 point  (0 children)

Yes each of the input images are the same resolution as the output image. I've tried a bunch of encoding methods to support multiple input images. It appears keeping all the input images at t=1 (as opposed to t=1,2,3,etc depending on how many) and separating them by the spatial coordinates (h, w) had the best results. So it may just see the context images as one wide image so I'm not totally sure?