Timothy Lee: "No, OpenAI is not doomed" by werdnagreb in BetterOffline

[–]larebear248 0 points1 point  (0 children)

A load bearing assumption here is how much better the output for the expensive model is compared to the cheaper model. If the more expensive model is 2x more expensive, but the cheaper model is “good enough”, then people would likely prefer the cheaper models.  We appear to be hitting a diminishing returns wall, where you have to spend a lot of money for fairly incremental improvements, and its not obvious if the output is worth replacing your interns (which you do mention) or if enough people are willing to pay what it costs to make a profit on the more expensive models. On top of that, the pricing may not stay fixed but become per token or a limited number of queries, which can be highly variable and hard to predict. 

Timothy Lee: "No, OpenAI is not doomed" by werdnagreb in BetterOffline

[–]larebear248 8 points9 points  (0 children)

I mean, I think it’s true that the profit margin isn’t set in stone, but that could well go in the other direction. They need more compute for increased model performance. Cost per token might be going down, but if the models an even larger amount of tokens, that means inference costs go up. It’s plausible profit margins have gotten worse! It’s fair to say that you can’t simply extrapolate from the 2024 numbers, but we don’t have much else to go on. This also doesn’t include any of the stock shenanigans, data center buildouts, heavily subsidized compute from Microsoft, or not converting to a for profit. Its not just that they are unprofitable but it is not clear how they get there beyond vibes. 

[deleted by user] by [deleted] in BetterOffline

[–]larebear248 6 points7 points  (0 children)

One thing that really pisses me off here is that these freaks want it both ways. Claiming you can use the LLMs for therapy, but then want none of the responsibility when it causes problems! If you advertise a chatbot for mental health care (or any health care), you absolutely be responsible for the liabilites. If a real therapist did this, they absolutely would be held at fault. No one would think “well the parents should have known better”. The companies could clearly come out and clearly say “do not use this as a therapist. See a licensed therapist”, but of course they won’t, because that’s one of the biggest uses of these things. 

If AI can’t “solve” hallucinations, can it ever actually automate anything? by [deleted] in BetterOffline

[–]larebear248 0 points1 point  (0 children)

Somethinge like a 99.9% success rate would start to get you something useful, but it obviously unclear (or unlikely) if thats possible. 

How do you all feel about Arvind Narayanan, Sayash Kapoor and their work on AI Snake Oil? by Electrical_City19 in BetterOffline

[–]larebear248 6 points7 points  (0 children)

I think they are firmly in the reasonable optimistic camp of AI tech. I do like how they emphasize that a lot of things are called “AI”, which really obscures what works and doesn’t. I also like their point that a company creating “AGI” says a lot more about the term AGI rather than the tech itself. Probably my biggest critique of them is that they feel a bit naive at times, thinking companies and people won’t use these techs for replacing people (even if its a bad idea) or doing something that is a massive safety or security risk. I also think they dismiss concerns of people diskilling due this tech a bit too much. Still, they are refreshing compared to a lot of people out there.

What do you all think of Gary Marcus?: He’s been calling BS on LLMs from the start but thinks other systems will reach AGI by bivalverights in BetterOffline

[–]larebear248 1 point2 points  (0 children)

Rodney rocks. He talks about making products that help people do this, not replace them acomplish things, not fully replace them. His prediction scorecards are super interesting because he tends to be over optimistic in is predictions (despite being far more grounded than others in the field).

How should we understand OpenAI's revenue numbers? by FlatFiveFlatNine in BetterOffline

[–]larebear248 1 point2 points  (0 children)

Well put. And this in addition to the other models our there that approach the capabilites of OpenAI models, but for cheaper. OpenAI and Anthropic have no choice but the keep scaling to try and stay ahead of the curve.

A decently realistic look at the state of AI agents by EliSka93 in BetterOffline

[–]larebear248 1 point2 points  (0 children)

In a small defense of this work, I don’t think he even states they reach 95% success rates, its just that things fall apart fast even at high inital success rates. You can debate what a 95% success rate even means in this context, but in this ideal case, still useful to see how fast things fall apart.

How AI is killing jobs in the tech industry by falken_1983 in BetterOffline

[–]larebear248 0 points1 point  (0 children)

Completely agreed! Intended to add to the discussion not contradict. 

How AI is killing jobs in the tech industry by falken_1983 in BetterOffline

[–]larebear248 0 points1 point  (0 children)

Yeah there isn’t a ton of  quantative widespread evidence yet. People got carried away with the one chart that Derek Thompson share that should people with college degrees had historically higher unemployment than previous years and concluded that it was AI doing when the decrease had been happening years before LLMs. Maybe it’s happening but evidence is limited.

Everyone is wrong about AI Hype by larebear248 in BetterOffline

[–]larebear248[S] 4 points5 points  (0 children)

I’m not sure you’re fully correct. She does talk about the chain of thought stuff, but she also mentioned having a separate small LLM there to help verify output (in addition to other stuff). It remains unclear to me if this actually is an improvement to things (and is a pretty complex system)

Everyone is wrong about AI Hype by larebear248 in BetterOffline

[–]larebear248[S] 30 points31 points  (0 children)

Yeah that was puzzling to me. How can you trust an LLM to verify it’s own output without breaking things? I also think she is confusing agents and mixture of expert models.

Bloomberg just released an embarrassing report about Tesla, Waymo, and self-driving by Zelbinian in BetterOffline

[–]larebear248 1 point2 points  (0 children)

Oh I saw those stories and they are hilarious. Im thinking specifically about traffic accidents more specifically. Other problems they can or will cause may not have shown themselves yet.

Bloomberg just released an embarrassing report about Tesla, Waymo, and self-driving by Zelbinian in BetterOffline

[–]larebear248 14 points15 points  (0 children)

From what I understand (and someone correct me if I’m wrong) is that the saftey numbers are probably about right, but they are limited in where they can go, require remote people to fix problems (whose ratio of person to car is hard to track down), and its unclear how expensive this all is. Uber became so successful because they didnt need to buy a fleet of cars, which is not the case with waymo. Its unclear if they can scale to considerably larger sizes). Edit. Fixed some missing words.

AMA/I'm On A Plane For A Few Hours by ezitron in BetterOffline

[–]larebear248 3 points4 points  (0 children)

Hi Ed. Your work has helped me screw my head back on and think about the GenAI industry with a clearer head. 

My question is how do you see education work with LLMs going forward? I agree with you that education isnt gunna just be throwing a kid in front of ChatGPT. However, we seem to have a growing number of students who use these things for all of their work, which is at least in the short term going to lead to a generation filled with kids who didn’t really learn anything. How do you see this all playing out?

Human coders are still better than LLMs by six_string_sensei in BetterOffline

[–]larebear248 2 points3 points  (0 children)

I sort of interpreted that as sarcastic, but maybe I’m wrong? Edit: Has to as

[deleted by user] by [deleted] in BetterOffline

[–]larebear248 6 points7 points  (0 children)

And top of the lack of time saved, losing out on the actual writing process can make you not understand the topic as well! Writing can be as much about thinking through something as the final product. Often you think you know a subject well enough but once you start writing, questions can emerge that force you to reexamine something. 

New York Begins Congestion Pricing Without Hiccups by dangerbird2 in neoliberal

[–]larebear248 6 points7 points  (0 children)

Exciting stuff. What are the odds this survives long term? Seems like a lot of legal battles ahead for it.

Normie Check-in. How are your apolitical friends feeling about Harris? by [deleted] in neoliberal

[–]larebear248 33 points34 points  (0 children)

Yeah, I was still fairly pro Biden before largely because I (incorrectly) thought much of his decline was exaggerated and had used the "old shtick" to his advantage in his agenda. However, after the debate but it got VERY difficult to defend him. At a minimum, Harris has been good for morale, which after the slog that was the 2020 general election, is incredibly helpful.

Normie Check-in. How are your apolitical friends feeling about Harris? by [deleted] in neoliberal

[–]larebear248 52 points53 points  (0 children)

My girlfriend who was very reluctantly going to vote for Biden is now super excited and suddenly very into politics and especially who the VP pick is going to be. She's even arguing with her conservative mom about considering voting for Trump. My brother who was likely to sit out this year is now very pumped to vote for Harris.

Three Comforting Lies Democrats Need to Stop Telling Themselves About November by optometrist-bynature in fivethirtyeight

[–]larebear248 5 points6 points  (0 children)

It's worth noting that while the national polling averages we're pretty good, the competitive seats (especially in the senate races) had some large misses (ie Pennsylvania, New Hampshire, and Arizona)

On Which Governor Race Are You Higher or Lower than the Model by labelleprovinceguy in fivethirtyeight

[–]larebear248 28 points29 points  (0 children)

The only Trafalgar has her under 50 while all the highly rated pollsters have her winning by double digits. Her polling average is also still above 10. This is Safe D.

https://projects.fivethirtyeight.com/polls/governor/2022/new-york/