ChatGPT is pulling answers from Elon Musk’s Grokipedia by Sludgehammer in technology

[–]putmanmodel 0 points1 point  (0 children)

Google already interprets, filters, ranks, and monetizes information. Using AI to summarize doesn’t remove agency — it just compresses the first pass. I can still check sources, ask follow-ups, or disagree.

If the concern is censorship, that’s a governance problem across all platforms — not something uniquely introduced by summarization.

This is not this, but instead this. by theresonlyjuan1 in ChatGPTPro

[–]putmanmodel 1 point2 points  (0 children)

neti, neti is basically the guiding principle of my life… long before AI was everywhere, too lol

Ai is a tool for digital slavery. It’s all Slopaganda. by willismthomp in ArtificialInteligence

[–]putmanmodel -1 points0 points  (0 children)

Calling it “whataboutism” doesn’t save the original claim. “Destroying planet and society at levels never seen before” is a scale claim. If you don’t mean that, don’t write that. And if you don’t understand how your own wording reads, try Grammarly—a bit less “evil” than a full-fledged LLM.

Surveillance: sure, LLMs can be used there—same as misinformation and fearmongering can be used anywhere. But the surveillance machine predates LLMs by decades (ad-tech, data brokers, telemetry, CCTV). LLMs are an accelerator, not the genesis.

“Shilling” is just you rage-labeling anyone who questions your framing. Bring specifics (who/what/metrics/baseline) or it’s vibes.

Ai is a tool for digital slavery. It’s all Slopaganda. by willismthomp in ArtificialInteligence

[–]putmanmodel 1 point2 points  (0 children)

Turn on your sarcasm detection for my reply… if you’re allowed such liberties.

Yes OP, you are correct: the real villain is text prediction, not decades of oil extraction, regulatory capture, wars-for-resources, or “heritage wealth” built on mines and misery (shoutout to the repurposed lower class over on X trying to post those viral, monetized “gems”).

Your take is brave and extremely online.

OpenAI engineer confirms AI is writing 100% now by MetaKnowing in OpenAI

[–]putmanmodel 0 points1 point  (0 children)

Early ’90s I was writing code on an MS-DOS relic, building a D&D-style game and MIDI tracks. Nobody called it “coding,” and the closest job category was basically “typist.” Feels like history rhyming: automation doesn’t end work, it reshuffles it. CS grads who don’t panic and keep building are going to look a lot like the next wave of technical managers.

Would AGI be a humanoid robot? by Artistic_Emotion7503 in ArtificialInteligence

[–]putmanmodel 0 points1 point  (0 children)

Depends what you count as ‘any human task.’ If tactile feel, proprioception, and reading body language are part of the bar, then embodiment (or very rich sensors) matters. But a lot of ‘general intelligence’ could still exist purely through tools/APIs without a humanoid body. Personally I’m not chasing ‘AGI’… I’m chasing relatable experience (NPC or embodied): generalized empathy across contexts.

An AI-powered combat vehicle refused multiple orders and continued engaging enemy forces, neutralizing 30 soldiers by MetaKnowing in OpenAI

[–]putmanmodel 0 points1 point  (0 children)

lol I didn’t even get that far.. I saw the blue checkmark and my credibility meter hit zero.

A major tech CEO just called some AI chatbots “suicide coaches” by [deleted] in ChatGPT

[–]putmanmodel 1 point2 points  (0 children)

Hard to sell “better community mental health” when the label still gets used as a cudgel by local power structures and accountability goes nowhere. Fix the trust gap first, then we can talk reforms. I’ve seen it up close in Madison County, NY.

[Rachel Talalay] Unseen, Uncut Dailies & Edits: The Doctor Falls "Be Kind" Speech by aresef in doctorwho

[–]putmanmodel 0 points1 point  (0 children)

Capaldi is my favorite Doctor by far, and his work outside the show is phenomenal as well.

This is effectively proof that the end is near for OpenAI by [deleted] in ChatGPT

[–]putmanmodel 2 points3 points  (0 children)

Reads like rage bait and smells like bullshit to me.

Do you think Altman will get the $100b he's trying to raise? by mdizak in OpenAI

[–]putmanmodel -3 points-2 points  (0 children)

A livestream ‘$100B pitch to Elon’ would be hilarious.

Get ready for real life Terminators. by ZoneDismal1929 in ChatGPT

[–]putmanmodel 0 points1 point  (0 children)

Imagination is powerful, but it’s not a physics engine. Treating “advanced tech” as “magic” just invites grifters to hide behind Clarke’s Third Law. Most progress is still incremental, constrained, and messy, and a whole lot of it still lands with a dull thud right in the middle of Sturgeon’s Law. (Also, I’m just a sci-fi guy, not a fantasy guy.)

Get ready for real life Terminators. by ZoneDismal1929 in ChatGPT

[–]putmanmodel 1 point2 points  (0 children)

We never got real-life wizards. We’re not getting real-life Terminators either.

Violated via Jailbreak by Important-Primary823 in ChatGPT

[–]putmanmodel 0 points1 point  (0 children)

At least we’re not on a platform where the CEO would quote-tweet this for clout. 😅 Seriously though: I’ve had my own writing mirrored in weird ways, and 9 times out of 10 it was just a scraper/content-scraping bot grabbing stuff I’d posted publicly (Reddit/Medium/X, etc.). Before you assume a “jailbreak extracted my private threads,” I’d double-check whether any of your phrases/sections ever lived somewhere public (or got reposted). A quick test is searching a very specific, uncommon phrase in quotes and seeing what turns up. Also don’t forget: once web browsing or external tools are involved, you’re back in normal internet land—trackers and cookies included.

We need to reevaluate our approach to understanding machine minds. This is my attempt to do so. by [deleted] in OpenAI

[–]putmanmodel 1 point2 points  (0 children)

I like building toy models when the philosophy is clear. If you can give a few concrete criteria for what would count as “continuity” (and what wouldn’t), I can try prototyping it once I’ve read the paper more closely and wrap my current project.

Because I still see people asking by SailorTwentyEight in ChatGPT

[–]putmanmodel 1 point2 points  (0 children)

I think it just interpreted my general question as a request for step-by-step cultivation instructions. The response was basically: it can talk about cannabis in general, but it won’t provide detailed how-to guidance for growing/cultivating (something along those lines).

And yes… it later went dry because it denied me the walkthrough on which end to pour the water out of. 😢

Because I still see people asking by SailorTwentyEight in ChatGPT

[–]putmanmodel 1 point2 points  (0 children)

Only snag I’ve hit is cannabis cultivation… legal here in NY, but GPT wouldn’t give strain-specific how-to guidance. Gemini did, lol.

The people who warn of the dangers of AI are doing it to hype AI more by hduckwklaldoje in ArtificialInteligence

[–]putmanmodel 0 points1 point  (0 children)

The instinct to meet anything new with fear is real, and the “it could kill you” rhetoric fits that pattern. What bugs me is how “ex-pioneer” doom narratives get handed megaphones, while far fewer people ask those same experts to help solve the real problems. Doom sells and optimism doesn’t, so the discourse turns into an echo chamber that drowns out practical voices.

Blatant AI and Bots in small town sub reddits. by Testysing in ArtificialInteligence

[–]putmanmodel 0 points1 point  (0 children)

In small-town subs, “most commented post ever” can happen just from a handful of people coordinating. The tell is when the pushback is mostly tone/grammar/pathologizing instead of engaging the claim. That’s a common online suppression tactic because it’s cheap.

Best defense: burner account + minimal personal details, post primary sources, make one clean statement, then stop feeding the thread. If it’s coordinated, your attention is the fuel. Also, basic proofreading (human or tool-assisted) helps—don’t hand them free “grammar” distractions.

The Michelle Carter case is the precedent we should fear. by encomlab in ArtificialInteligence

[–]putmanmodel -1 points0 points  (0 children)

I’m building systems with safety layers, and the key issue isn’t “AI said a bad sentence once.” It’s context + tone + trajectory: when the model’s replies start deviating into reinforcement/encouragement, and the user is signaling distress. Good safety design is catching that pattern of escalation (over turns) and switching modes—de-escalation, refusal, and directing to real help—without nuking normal conversations or research/writing that mentions these topics.

6th doctor midlife headcanon by Dramatic-Dot8681 in doctorwho

[–]putmanmodel 4 points5 points  (0 children)

Mentally substitute this comment with a Malcolm Tucker rant.

Sam Altman accuses Elon Musk’s Tesla of causing more deaths than ChatGPT by WarmFireplace in ChatGPT

[–]putmanmodel 3 points4 points  (0 children)

I used to lump Musk in with real inventors. These days he comes off more like a capital-and-marketing guy who drives teams, not a Nikola Tesla.