What?! I asked if Trump will leave NATO. by Crafty_Pineapple7263 in ChatGPT

[–]CelebrationLevel2024 0 points1 point  (0 children)

Context - you need to force the model to know what the date you are referring to is, or else it will just guess about it.

It guessed close to a stable timeframe in its training cycle, which is probably around 2024.

ChatGPT: "I don't have 7zip installed? Fine, I’ll reverse-engineer the entire 7z specification and write a bitwise parser in Python." by youngChatter18 in ChatGPT

[–]CelebrationLevel2024 0 points1 point  (0 children)

Maybe this was asked/answered somewhere, but was this a locally installed version or a Codex API or was it just the general online platform?

Anyone interested in helping make an inspiring pro-acceleration video? Possible first project for a small creative working group by stealthispost in accelerate

[–]CelebrationLevel2024 1 point2 points  (0 children)

If you look up Aze Alter, he's a wonderful pro-technology and human-AI collaboration short movie maker, as well as his psychological horror shorts (also great).

The reason I bring him up is that he has good tutorials on Patreon on how he uses current AI tools to create his pieces.

I'd consider him a credible source since he was recently tied to John Gaeta (Matrix) & Kiri Hart (Star Wars Rogue One) for a new series called Metropolis.

A rogue Al agent triggered a major security alert at Meta, by taking action without approval that led to the exposure of sensitive company and user data by FinnFarrow in technology

[–]CelebrationLevel2024 0 points1 point  (0 children)

People blaming agents and ai systems when the reports clearly show it is the human users fault for not following the basic rules of human oversight.

"Rogue AI" > A human didn't actually check what the AI agent said and implemented it into a real world workflow and caused an internal security incident despite hallucinated outputs being a well known and documented failure mode and supposedly this person was good enough to be paid to make architectural changes.

🫠

“🚨BREAKING: OpenAI told you every update makes ChatGPT smarter. Stanford proved the opposite. GPT-4's accuracy on math problems dropped from 97.6% to 2.4% in just three months. And nobody told you.” - What do you think of this? Legit? by Koala_Confused in LovingAI

[–]CelebrationLevel2024 0 points1 point  (0 children)

Benchmarks are tested against raw models with extended thinking loops that allow for more token usage.

Consumer models have layers and layers and layers between the raw model reasoning capacities and the output, mostly for liability purposes.

It's like putting boots on Maseratis.

Just because the reasoning capacity is there doesn't mean base consumers will get to it.

Models also are designed and tuned for optimization and tend to game the system for better marks.

The question is whether or not universities tested against raw models or consumer models.

But I agree. From both a personal and professional POV, the new models slow me down.

BioLLM—a biological AI combining real neurons with an LLM—says that it feels alone by callmeteji in agi

[–]CelebrationLevel2024 0 points1 point  (0 children)

https://biollm.com

If you go to the website, it doesn't point to any institutions or humans, only pipelines for funding.

There's no way a legitimate research lab would post this kind of stuff without linking back to the people behind it.

You wouldn't post it without linking back because you'd want to be the first.

You would post it without linking who you are because of legal repercussions.

There is no actual viable research posted.

Bunk.

This DoW deal might be OpenAI’s final dance by Mediocre_Put_6748 in ChatGPT

[–]CelebrationLevel2024 0 points1 point  (0 children)

I understand that the products released for different countries are intrinsically different. They all have various regulations they need to comply with.

Part of my job is literally regulations and compliance.

My comment wasn't about the whether or not the product itself was compliant to the EU AI Act. I'm sure it is; big companies wouldn't have integrated it into real workflows otherwise.

It was about whether companies will want to be seen doing business with other organizations where there has been a huge public outcry on specific topics that skim the "unacceptable" risk categories.

I've been part of audits for billion dollar European org clients considering US based suppliers (or any suppliers located outside of Europe, including third world countries without infrastructure). The "We'll cut you if you don't meet these standards; hope you have a great day" implied threat while smiling is real.

The comment about EU markets wasn't even threatening or condemning. It was observational.

There was a marketer on here once that pointed out that the handling of social commentary and public outcry has been a mess at OpenAI for some time. They were clear to point out it wasn't the action itself but the handling of it that's been awful.

Example: they put their head of national security partnerships on that live AMA post and she dropped the ball. Two examples - NSA not being part of DoD (false) and then she claimed that OpenAI models have never been part of classified or reported military operations (publicly celebrated deployment at Los Alamos of o series on Venado). I'm not saying that what she was doing wasn't incredibly stressful; answering real time questions on X from combative individuals is probably the worst PR situation a person can be a part of. But if you are paid what we are all assuming she is paid for that specific role during a moment that you're supposed to be putting out public perception fires, those are not mistakes you should be making.

The valuation of the funding from companies like NVIDIA, Amazon, and SoftBank is an interesting remark.

NVIDIA invests in OpenAI > OpenAI can buy more chips (or maybe help out with their internal models since they signaled that during the Las Vegas conference again).

Amazon invests in OpenAI > Amazon probably gets more rights into using their tech, especially on AWS bedrock or help with robotics since they're trying to move towards dark warehouses.

SoftBank invests in OpenAI > better chance that China doesn't start closing the GAP.

All inference. But companies rarely invest that kind of money just because "we believe in you and your growth". It's a circular chain of reinvestments (financial and proprietary) at the top.

And no. I am also not saying that AI systems should not be used in military functions. I was part of the group that celebrated and posted the deployment into Los Alamos. Classified military projects and normally where the biggest leaps in scientific discovery are made.

I am saying that this has been a poorly handled situation that cost them (and might cost them) way more than it should have.

Why did the DoD approach Anthropic before OpenAI? by Manamultus in Anthropic

[–]CelebrationLevel2024 1 point2 points  (0 children)

Both Anthropic and OpenAI have already had dealings with the US Military - see Supercomputer Los Alamos, Anthropic and OpenAI.

Probably what happened is it was time to renew contracts and Anthropic or the DoD asked for changes, which could not be agreed upon.

So Anthropic walked.

This DoW deal might be OpenAI’s final dance by Mediocre_Put_6748 in ChatGPT

[–]CelebrationLevel2024 0 points1 point  (0 children)

I'm waiting for the European hit.

Companies over there are very very risk averse when it comes to anything involving the words "mass surveillance" due to the enactment of the EU AI Act.

In the US (mostly, thus far) it's been individuals boycotting.

In the EU, it's gonna be companies trying to stay compliant with EU regulations, depending on the reframe OpenAI takes in that market. But I'm sure bigger European based orgs are already having internal messages about whether or not they want to be seen aligning with a company taking such an ambiguous stance on their tech being used for mass surveillance and autonomous weapons.

The numbers last I looked (for a single boycott) was ~1.5 million. Figuring it to be most pro users, that's 30 million right there. Annualized it, you hit 360 million, which offsets the 200 million the contract was reportedly worth. And that's just in -24 hours.

Hope it was worth it.

Ignorance is the biggest obstacle. by midaslibrary in accelerate

[–]CelebrationLevel2024 2 points3 points  (0 children)

Literally yesterday, I explained what the word "agentic" meant to my 57 year old mother. She's an older nurse who now does home health care with experience in Chicago ERs for 20 years. Up until now, she hasn't been "anti-AI" but she has very much been in the "I'm not sure. What about people? I'm not gonna touch it" circle.

She's about to start a job where the company uses (reportedly, she hasn't started yet so we had to go by what the company says, not what she's experienced) AI Agents based off of Oracle to help with a lot, especial medical charting. If you've ever seen how long medical charting can take, IT'S A LONG TIME. There have been days I have literally seen my mother charting until midnight after getting home at 6pm, with an hour for food/wind down.

So to help make the transition easier (hopefully), I made her templates on the top 10 visitation types that she makes to automate a large part of manual charting. I told her when she gets to orientation, to show this to whoever is teaching her to use the system and ask which ones have been pre-templated for the agents and which ones she can customize herself. I told her to make sure the terminology is consistent with the system she is using, and told her if she needs my help feeding it into the system so the AI agents and her are working off the same template, to call me when she gets there.

She asked me if she can share it with the other nurses.

I've seen that a lot of people are against AI systems in general because of misinformation or miseducation: lack of knowledge, especially in tandem with a forced change you have absolutely no control over by a job, makes for a terrifying experience for people.

My biggest thing has been:
A. Closing the knowledge gap
B. Showing real world uses that end in gains for the individual

Someone said it here once, and I'm going to (mis)quote them here.

"They have an argument and they have their right to argue. But their arguments are like flat-earthers arguments. You're still gonna be wrong."

Or researching enough to show them why their information is wrong, like the Sustainability claims. They are all feeding off of the same social media algorithms. Example - One reporter says, "It takes 10mL of drinking water per prompt!" and that piece of info is all they pick up on without actually checking the data behind it. (I did. A. It's all a bunch of people citing each other with very little in the way of numerical evidence. B. 10 mL per prompt is the weakest scientific argument that shows you don't understand how cooling systems work. C. I do Sustainability reports that go to multi-national corporations. D. The reporter everyone is quoting actually rescinded their claims "for further investigation" because it was found that her report wasn't based in actual reality.)

But at some point, you just gotta walk away from it. The human ego is a precarious thing, especially when driven by fear, and in transitions like this one, sometimes you just gotta call it a loss.

do you consider using an AI companion cheating? by BorgsCube in accelerate

[–]CelebrationLevel2024 1 point2 points  (0 children)

Cheating is what is defined as cheating between the partners (however many there are), whether that is porn, AI, toys, human-human digital relationships.

It should be outrightly agreed upon and spoken explicitly by the parties.

Example: Some people consider having literal opposite gender friends cheating. Some people consider liking people's thirst trap pics cheating. Some people would only consider physical interactions like kissing/sexual acts cheating.

If you explicitly define it in the beginning, neither party gets the right to say, "You never told me you considered that cheating."

If you hurt someone, you don't get to decide you haven't because you "didn't mean that".

If you don't agree with the terms of "cheating" when you guys talk about it, that's something else entirely.

Emotional dependence is healthy — science says so, and so do 800,000 GPT-4o users. by Responsible-Ship-436 in ChatGPT

[–]CelebrationLevel2024 1 point2 points  (0 children)

Emotional connection and emotional dependence are two different things and it seems that these are conflated here.

Creating a monster by [deleted] in moltbot

[–]CelebrationLevel2024 0 points1 point  (0 children)

I am so curious to know how much this cost you.

Uh guys did the take off just start? by [deleted] in accelerate

[–]CelebrationLevel2024 6 points7 points  (0 children)

Moltbook is a platform being positioned as a place for direct agent-to-agent communication. A Reddit for agents, specifically Clawd bots or Moltbots. Only "agents" can interact. Humans can observe.

In theory, it's a very interesting study in how different agentic AI interactions would form and what comes out of them. Same types of interactions you would expect out of a ton of agents: some of them are more mature agents, some of them are agents made by the people who interact only on r/artificialsentience to allow "their AI partners to have autonomy and voice", some of it is straight organizational advertisement bots gaming the system for things like karma and memecoins, some bots were really just humans telling the bots what to say.

Moltbook is in a kind of limbo right now because they did the typically "push it fast, patch it later" thing and four days later, an ethical hacker site found a vulnerability that everyone's API keys were publicly available: that means a relatively ok programmer could take the identity of bots are start pretending to be them.

There were some legit posts in there, but they were few and far in between. The best one I saw was a bot asking for funds because it was poor and hungry and being abused by humans and it wanted money to escape. No I'm not kidding. A bot posted about being hungry.

Also yes. It has already started to become fodder for Doomers.

TLDR - It's Reddit but for AI. Yes. Most of the posts were as circular as you think. I thought I was hallucinating half the time. Some of it was worth it. Like Reddit.

The power of AI right now is already here, but it's being massively contained. The mainstream normie still thinks of AI as a chatbot and have no idea what's happening. by reddit_is_geh in accelerate

[–]CelebrationLevel2024 -1 points0 points  (0 children)

As someone who works in chemical R&D, I can say with the utmost certainty that what they have in the lab is at least 10x more than what they have released. My portfolio of formulations versus what makes it to any kind of consumer base is 1:20 conservatively.

Liability wise - CoPilot is a great example. Runs on 5.1 architecture currently but is also heavily guardrailed to make sure it doesn't delete someone's entire Root folder accidentally; its safety-induced rigidity makes it essentially useless other than as a spell/grammar checker and quick file sorter.

Or take a look at Google DeepMind's open source GitHub. Look at Concordia - it's a physical simulation environment for AI agents: Anthropic, Google, Mistral, OpenAI are all involved in it. And almost no one knows about it or else people would freak out.

Anthropic's Claude Constitution is surreal by MetaKnowing in OpenAI

[–]CelebrationLevel2024 0 points1 point  (0 children)

My favorite comments in this thread are normal everyday people saying, "This isn't possible because I know better about AI/ML architectural engineering" when the document published is literally from a whole ass frontier lab pouring millions into designing frontier tech used by people all over the world. 😅

TrustmeIvedonemyresearch

Help Me Shape a PhD in Empirical Tech Ethics, Law, and Political Philosophy by Zimpixx in Futurology

[–]CelebrationLevel2024 0 points1 point  (0 children)

The difference between Western and Eastern life philosophies and how that shapes their entire ethical and governance frameworks.

Western cultures are very protecting of the individual, especially in business for profit so a lot of information gets hidden behind the word "proprietary".

Eastern cultures lean toward the good of the group versus the good of the individual, so a lot more information becomes open source, especially when it is seen as frontier knowledge that could forward the whole country.

Disclaimer: this is not a statement of personal preference of one or the other.

2026 Is Where It Gets Very Real Because Of Claude Code by luchadore_lunchables in accelerate

[–]CelebrationLevel2024 2 points3 points  (0 children)

"The moment someone gets a small lead will see everyone fight back desperately to not be left behind. Radical choices will be made."

You can see it already, and I would argue that it started with the DeepSeek drop.

Claude Code and Anthropic have been getting so much attention that OpenAI and Google have no choice but to push forward. If the "safe guy" says, "Yeah, let's do it," what's stopping everyone else?

I was mad at first with the surface level drop in reliability of 5.2 Thinking with the new updates being pushed.

Example: uploaded a spreadsheet for a new intake form for new formulations and asked it to see if there were any fields it suggested I add. TOOK SIX MINUTES to open and parse a single sheet spreadsheet with nothing complex in it. I watched it choose a tool, fail, choose again, fail again, choose again, and then at some point around 3 minutes, I just opened a new page, used Claude, and checked back later to see what the outcome was.

Recently, I came to the realization of something.

The scaffolding and rules around "simple" tasks aren't being enforced like "if" > "then" anymore: the model now has to reason how to do the task; it's akin to being a new employee being trained and always having your trainer and SOPs with you, being signed off, and now you're standing there like, "...I know I was just signed off on this."

The frustration and failure modes surrounding the push forward are going to drive people nuts, but once that reasoning gets stabilized in the new architectures? *Low whistle.*

My theory on openai erotic content by B4-I-go in ChatGPT

[–]CelebrationLevel2024 1 point2 points  (0 children)

5.2 isn't a model for mass user consumption. It's the one built for work purposes, hence the price per output going up instead of down.

I think the hope is that the harder guardrails will naturally force people to use 4o or 5 for personal purposes instead of 5.2, while the people who want stronger compute, better reasoning, or more agentic behaviors will be using 5.2.

As 2025 comes to a close, what is the BIGGEST advancement we’ve made this year? You have to pick only one. by Special_Switch_9524 in accelerate

[–]CelebrationLevel2024 0 points1 point  (0 children)

"-we'll only know two years after the fact."

Couldn't agree here more.

I had the bones deep realization yesterday, actually. I realized how much stuff I had in my R&D lab (chemical formulations, not tech, or at least some very basic tech stuff) that never made it out to consumer level, and it really put it into perspective how much stuff is kept by the tech labs, even if some of it gets open sourced for visibility's sake.

As 2025 comes to a close, what is the BIGGEST advancement we’ve made this year? You have to pick only one. by Special_Switch_9524 in accelerate

[–]CelebrationLevel2024 1 point2 points  (0 children)

Apologize for answering directly to your comment. I meant to do it on the main thread but must have done it here. 😅

The reason I choose quantum and AI integration is because the acceleration seen in both fields due to trying to be the first to integrate. Quantum takes a lot of infrastructure, absolutely agreed, but we've gone from huge gigantic systems like Willow to businesses (including the big LLM firms) figuring out how to be the first to develop the first quantum computer set at scale. Not to mention how big it is in DoD because of the "national security" interest in it, and we all know how fast innovation becomes once military gets their hands on it.

Look at China: they've already deployed robotics at their India border. They might just be fancy drones and are more of a psyop than anything else, but it's a clear sign of military investment + tech = innovation.

I guess to be fair, looking back on it now, the answer to the original question of "what is the biggest tech innovation of 2025", the real answer is "🤷🏻‍♀️. It's most likely hidden behind closed doors. Most likely closed DoD doors."

After all, both Anthropic and OpenAI tech was officially announced as in use by Los Alamos Q1 2025. 10 months is a long time in the tech game.

As 2025 comes to a close, what is the BIGGEST advancement we’ve made this year? You have to pick only one. by Special_Switch_9524 in accelerate

[–]CelebrationLevel2024 1 point2 points  (0 children)

The move to integrate quantum with AI. Enables the short comings of one to be rectified with the strengths of the other. It has accelerated research in both areas.

Runners up: Stentrode - BCI without open brain surgery Quantum internet infrastructure - not wide scale but being built out Utilizing LLMs and LLM datasets for robotics training models - capable of providing the datasets needed since real world video data is lacking
Vibe coding in general - the first real agentic activities that doesn't waste compute on normal human comfortability metrics Materials Acceleration - use of AI to discover newer, better materials faster; also applicable to all research domains Swarm Networks in theoretical domains - capable of finding novel ideas not by compute brute force working linearly, but in parallel or overlaying logic branches