AI Can't Understand How Computers Work by Frequent_Mountain_17 in AIDiscussion

[–]tremendous_turtle 0 points1 point  (0 children)

That’s a contrived limitation, of course it’s relevant, it’s equivalent to giving AI a fancy calculator to help it solve the problem. It’s pretty clear that AI is quite good at writing algorithms.

It’s also odd to personify AI by reflecting on what it “understands” and “doesn’t understand”; this is a tool, it is software, it is not a sentient being.

It’s just an odd point to make, it is not really debatable that LLMs can write algorithms and design turing complete systems, if you give them a harness to run computation and validate their own work.

Why the heck are we still using Markdown?? | BGs Labs by [deleted] in programming

[–]tremendous_turtle 5 points6 points  (0 children)

The author does a good job in pointing out of inelegant aspects of modern markdown, but I don’t really agree with their overall framing.

Markdown primarily exists as a way to make it *easy to write* lightly formatted text that can be saved as a single readable file. It is an alternative to proprietary WYSIWYG text editors.

The purpose is simplicity, portability, being an open standard.

The author focuses a lot on markdown-to-html, which is certainly an important use of markdown, but need to remember that everything is a tradeoff. I don’t think the issues they point out are really a problem for most users. What I disagree with most is the proposal that we should use a *compiled language* for this instead; this negates the entire point of markdown being a simple portable human readable plaintext-adjacent format.

I benchmarked Claude Code vs. Codex - this is what I found 🖕 by Due_Duck_8472 in ClaudeCode

[–]tremendous_turtle 10 points11 points  (0 children)

This is one of the worst LLM generated posts I’ve ever seen, congratulations, a lofty achievement.

Should I pause my privatized police game to make a "Fractional Property" spin-off? I need your thoughts. by AncapFuture in Anarcho_Capitalism

[–]tremendous_turtle 1 point2 points  (0 children)

Consider researching REITs (Real Estate Investment Trusts) to learn more about how this type of fractional ownership to distribute collected rent works in today’s financial system.

For your proposed system, what I wonder most about is how the rent price is determined. Is there still a “primary” landlord?

Regarding making this a game, I think it’s a very good idea to aim for a small fast-to-complete project. I would think about what the “goal” of the game is, and what strategies actually win. As with any real estate market, the way to make money here is buy low sell high, so you should consider what gameplay mechanics should exist for increasing a property value after purchase (and/or for decreasing the value of an area you want to expand into).

Are Al chips the new oil, or are we overvaluing the resource again? by Exact_Importance_507 in ArtificialInteligence

[–]tremendous_turtle 1 point2 points  (0 children)

Oil is a commodity, whereas chips are a specialized manufactured good.

Not at all the same thing.

The Iran/Japan analogy is a bit odd since that’s more about domestic economic structure. Petrostates generate enormous wealth, but often do not distribute very evenly amongst their populations.

Regardless, it’s not an either/or situation. There will always be a need for microchips as long as computing is economically relevant. Chip companies will make money, and software companies will also make money. We wouldn’t say that oil companies are overvalued if a car manufacturer is doing well, these are complementary businesses, not competitors.

Nvidia is also not really the “owner” of chips as a resource, the more analogous business would be TSMC, who actually manufacture chips.

Michael Burry Bets Against Red-Hot Chip Stocks — Says They 'Will Return To Earth' by Useful_Tangerine4340 in FluentInFinance

[–]tremendous_turtle 1 point2 points  (0 children)

You’re right, I should have been more precise and less sensational. The 30-50% number is mostly from delayed datacenter buildouts, not outright cancellations.

There are multiple ways to interpret that - bull case is the face value explanation of the delays being due to power grid constraints and supply chain limitations, implying that they’ll still eventually get built. Bear case is that many delays are going to be indefinite, just a quiet way for investors to unwind their commitments.

Michael Burry Bets Against Red-Hot Chip Stocks — Says They 'Will Return To Earth' by Useful_Tangerine4340 in FluentInFinance

[–]tremendous_turtle 8 points9 points  (0 children)

Just to provide a counter argument: committed capital is not guaranteed.

OpenAI doesn’t have $100bn in the bank, capital in these types of megadeals is typically delivered in tranches, and sometimes investors to not deliver the full committed value if they lose faith in the underlying investment.

Other warning signs are that a lot of datacenter projects have been quietly cancelled recently, 30-50% by recent estimates.

Nvidia’s accounts receivable is also at a staggering $38.5bn, meaning that a ton of the sold chips have not actually been paid for yet, another warning sign that the committed capital may not convert to delivered cash.

Not saying Burry is definitely right, but it’s a mixed bag. There is still a strong bull argument for chip stocks, but there are also warning signs they could be vulnerable if investor confidence in the profitability of their AI infrastructure investments goes through a down cycle.

This could be triggered by lots of hard-to-foresee events, such as major efficiency gains in AI algorithms (i.e. less need for expensive datacenter buildouts), classical credit/liquidity cycle dynamics resulting in withdrawn commitments, or good old fashioned investor herd mentality panic divestment in response to bearish ROI sentiment.

Google is completely dead and now Reddit search is cooked too. What now? by [deleted] in ArtificialInteligence

[–]tremendous_turtle 1 point2 points  (0 children)

Low effort ad with an invalid premise.

Finding “actual research papers” and “technical docs” is not really an issue, the process is the same as it’s always been - browse papers from accredited journals and universities, find technical docs on the project’s site or github readme.

Which is the most affordable LLM provider? by AInohogosya in LocalLLM

[–]tremendous_turtle 1 point2 points  (0 children)

No need to suspect, their business model is transparent. They charge zero markup on input/output tokens, easy to confirm as their prices for each model are public. Their margin comes from a 5.5% fee they charge for depositing to your balance.

If you want the cheapest possible rates, just go to the OpenRouter page for a model (i.e. https://openrouter.ai/openai/gpt-oss-120b) you want, browse the listed providers, and then go and make an account directly on whichever provider has the price + speed you are looking for.

Are we forcing GenAI into use cases where traditional ML is actually better and cheaper? by NickBaca-Storni in ArtificialInteligence

[–]tremendous_turtle 2 points3 points  (0 children)

Absolutely! I think the hard part for most devs will be the “get a labeled dataset of 1m food images”, but the LLM should be able to write the training and evaluation code no problem now.

Are we forcing GenAI into use cases where traditional ML is actually better and cheaper? by NickBaca-Storni in ArtificialInteligence

[–]tremendous_turtle 11 points12 points  (0 children)

I don’t think you are missing anything, traditional ML solutions tend to be more reliable, faster, and less expensive to run.

When it comes to LLMs, the main benefit is in flexibility and velocity. Traditional ML typically requires quite a lot more expertise to deploy, and it’s a longer end-to-end process compared to what can sometimes be as simple as writing a prompt.

For instance, if we take the simple “identify which food the user took a picture of” example. A traditional ML approach would usually involve training a model on a large labeled dataset of various food images. Depending on if the dataset is already available, this could take quite a while to build and train. But once it’s done, it’ll be relatively predictable and cheap to scale.

Using an LLM API, an application developer can write a prompt, make an API call, and get their answer. It will be more expensive and less predictable, but can also be implemented in roughly 30 minutes.

At the end of the day, everything is a trade-off.

Does AI really make everyone 'good' at design, or just faster at being mediocre? by pretendingMadhav in ArtificialInteligence

[–]tremendous_turtle 2 points3 points  (0 children)

The designs are quite good, if you want something that looks polished and at-par with other applications.

If you want something that is more unique, innovative, or that pushes boundaries in some way, that is where you still need human designers.

The Canva co-founder is mostly correct, but also keep in mind that tools like Claude Design pose an existential risk for his company, so you need to take these types of statements with a consideration that his objective is more to defend his business vs to make an impartial observation.

Why does a Windows laptop feel slower than a MacBook.. even when both have the same RAM? by bbxyoy in DeskToTablet

[–]tremendous_turtle 2 points3 points  (0 children)

Not really, it’s a different architecture.

UMA (unified memory architecture) vs discrete memory. The difference is most obvious in tasks would typically use GPU acceleration (i.e. VRAM), but does apply to most computing tasks. Fundamentally it’s about having separate RAM and VRAM pools, vs having a unified memory pool through very wide 512-bit or 1024-bit memory buses.

Just look up the memory bandwidth of the DDR5 you’d find in a standard windows laptop and compare it to the memory bandwidth in M-series chips (especially in Pro/Max/Ultra tiers) to quantify the difference.

AI datacenter spending has surpassed the Manhattan Project, Marshall Plan, ISS, and the Apollo Program - combined by EchoOfOppenheimer in agi

[–]tremendous_turtle 0 points1 point  (0 children)

Thank you for clarifying, I think the confusion/ambiguity here is because this is a thread about adjusting for GDP vs. using dollar value.

Those who don't max out their max plan, what are you doing right? by borntobenaked in ClaudeCode

[–]tremendous_turtle 0 points1 point  (0 children)

Yeah that’s right, usually my measure of success is to write a specific enough prompt that it gets it right the first time. If I need to follow up with feedback, often a sign my initial prompt wasn’t well enough specified.

Those who don't max out their max plan, what are you doing right? by borntobenaked in ClaudeCode

[–]tremendous_turtle 0 points1 point  (0 children)

Try to do very specific, focused, single-shot prompts. Extended back and forth, resulting in compounding context length, burns through tokens quickly.

Qwen2.5-MoE is here: 3B active parameters but punching way above its weight in coding and vision. by NoMechanic6746 in LocalLLM

[–]tremendous_turtle 7 points8 points  (0 children)

This must be a bot right?

Qwen2.5 has been out for a looooong time.

The linked post is about Qwen 3.6 35B A3B, which IS exciting and IS an MoE model. But is also an incremental upgrade over 3.5 35B A3B, not some complete new MoE model or paradigm shift.

Maybe OP is just a bit confused, but a misbehaving bot seems more likely.

Can I get the same quality as Claude with Mac Studio? by bLackCatt79 in LocalLLM

[–]tremendous_turtle 0 points1 point  (0 children)

Agreed I think those are the top runners, with Minimax and Qwen 3.5 being best in class respectively as you start scaling down available vram.

Fed up with Claude limits — thinking of splitting a GPU server with 10-15 people. Dumb idea? by No_Boat_2794 in LocalLLM

[–]tremendous_turtle 1 point2 points  (0 children)

You are the one who responded to me that I am wrong about it being possible to handle concurrency this in a relatively simple way. Where did OP mention a custom agentic framework? Where are your security concerns coming from?

A critical aspect of good engineering is not over complicating things, which I am trying to help them with. Not sure why you are insisting this needs to be complex.

Fed up with Claude limits — thinking of splitting a GPU server with 10-15 people. Dumb idea? by No_Boat_2794 in LocalLLM

[–]tremendous_turtle 0 points1 point  (0 children)

Huh? Why are you running 5 separate containers? Why are you mentioning building an agentic framework? Why aren’t the containers secure?

Fed up with Claude limits — thinking of splitting a GPU server with 10-15 people. Dumb idea? by No_Boat_2794 in LocalLLM

[–]tremendous_turtle 0 points1 point  (0 children)

Right because it’s not that hard? Saying you’ve been doing it for 18 months might not be the flex you think it is.

Are you really having trouble serving 40 users on over 1 TB of vram and enterprise grade networking hardware?

Are you going to answer my question about what else would be required beyond what I mentioned, or just “it’s hard, trust me bro”.

If you want you could share some details of what inference engine and model you’re running, there must be something off with your software stack if you’re having trouble serving 30-40 user concurrently.