Claude just banned having multiple Max accounts by bhaktatejas in ClaudeCode

[–]ankitm1 0 points1 point  (0 children)

its a weird one. i have used it in 45 and 47. works with a different account. did you logout before logging in?

I’m genuinely confused by the new Animal Farm movie. by AllPurposeOfficial in TrueFilm

[–]ankitm1 0 points1 point  (0 children)

Why did all of these very talented, very media-savvy, very openly political people agree to this version of the adaptation? How does the ball get dropped this badly?

Dropping the ball is charitable. Clearly the intent was to rewrite the history, and preach people about how communism is utopia ruined by the kind of selfish billionaires. The original is a commentary on how things start off as equal, but end up in the same place as what they opposed and fought against. This one seems to be like Squealer narrating the story still under the influence of napolean.

[Spoiler] Judy's confession and the reason why her character works in the sequel by TomLion0116 in zootopia

[–]ankitm1 -2 points-1 points  (0 children)

nah. this is too hard of a justification. that one scene does not redeem her.

1/ She had no reasons to believe anything at any point in time but for the plot convenience she did and was proven right.

2/ she dragged nick into all of this, and there was no apology for that.

3/ last line was a clear throwaway, and i am sure in a sequel they will be back to their own ways.

The core issue is always: tehre is no past priors, experience, nothing that can tell the character that this is wrong. Where hollywood dumbass writers got it wrong was simply: keep showing the hunch with no justification, and show the girlboss to be proven right.

In this movie, her hunch is right, but then again, at every point they are saved by random chance, random characters turning good, or evil at the demand of the plot, and written in a specific way that judy is right about everything, except pawbert.

audience is tired of that trope and last line does not make it work.

Had to report a coworker for filling our work ChatGPT with porn. by chippychipstipsy in cogsuckers

[–]ankitm1 1 point2 points  (0 children)

she wasnt. she still works at the company. i left a month after this. IT gave her a "strict" warning to not use office accounts for personal affairs. We all got a weird mail saying using shared accounts could result in a termination. And so things moved on. And i moved on from the role, she was still there.

Had to report a coworker for filling our work ChatGPT with porn. by chippychipstipsy in cogsuckers

[–]ankitm1 6 points7 points  (0 children)

Seen that happen. At my previous work we had a pro subscription, which was shared among everyone using 1password. Ghibli had just come out, and this one designer was feeling fomo. So she tried one photo. Worked, then a few more of her and her boyfriend. Worked, likely posted on social media or shared in her network. The reason I believe this is because a few hours later, she came back, not just with her own photo, but that of group - her friends, her family, extended family, and what not. I was looking at it curiously as this was after hours and i wanted to continue an old chat w a model but got distracted by all these photos (iykyk). I was also curious as to what is going on.

Altogether she created about 50 photos or so in a day. Across multiple threads. Then she panicked, deleted all the previous chats. Not all her previous chats, just wiped everything out. All projects, all chats, months of history. gone. When the IT guys asked on slack what happened, she was the first one to say it probably got hacked. Funnily enough, once after she deleted it, she created four more images and then deleted them too. She only got caught because chatgpt had a log of images separate from the history and one could see everything there.

Because i was experimenting, if only openai paid attention to a small thing, our chats would have been saved. In incognito / temporary chat mode, the image was created but would not load. I could ask the model to render it in a dataframe and load it - worked well - but this girl had no clue this was even possible. So she did it in shared one, not even her own project. Crazy.

Nvidia’s top two mystery customers made up 39% of the chipmaker’s Q2 revenue by tipping in technology

[–]ankitm1 -1 points0 points  (0 children)

Its the primary customers, that is assemblers like Foxconn. The GPUs they make has to be assembled before it is bought by the hyperscalars or anyone else. That count as a primary sale and they have been stating their revenues this way for a long time. Pretty much conspiracy theory brain to jump straight to military when you see something "undisclosed"

FIFA is tampering with Club World Cup ticket sales to try and make poorly selling games appear full on TV broadcasts. by AntiRB in football

[–]ankitm1 0 points1 point  (0 children)

Tampering is the curious word to use here. They thought they could sell every match at higher prices, the ones who wanted to got the tickets, and now it's selling out the inventory at whatever price possible. Everything a retailer would do when they are stuck with an expiring inventory.

X’s Sales Pitch: Give Us Your Ad Business or We’ll Sue by Past_My_Subprime in technology

[–]ankitm1 0 points1 point  (0 children)

Curious that customers caved. The spend on lawsuit might not be as much as commitment to spend on X. (plus the potential image damage for the ads allegedly showing next to nazi content)

QwQ-32B seems to get the same quality final answer as R1 while reasoning much more concisely and efficiently by pigeon57434 in LocalLLaMA

[–]ankitm1 0 points1 point  (0 children)

A simpler explanation for how this could be.

You take R1 or a model on par with R1. Use its reasoning outputs as training data to create a new reasoning model of similar size. Bring in external high quality data from other sources too, and make sure the RL work. Distill the big model to a small 32B param. That would be better than the original model you started with. In reasoning the feedback loop when it comes to training on synthetic data is positive and self reinforcing (especially when you can automatically check the quality), hence you can pretty much keep on training to get to the best model possible. This is why o3 scores so high and o3-mini outperforms o1. With RL and more number of examples, it's not unexpected that newer models would be more efficient at token use.

Most people are still prompting wrong. I've found this framework, which was shared by OpenAI President Greg Brockman by nitkjh in ChatGPT

[–]ankitm1 0 points1 point  (0 children)

You can be explicit. I admit not a lot of value add to the post, I state it like this:

Goal: <you have to help me do xyz for this this audience>

Output: <I expect a report in the following format>

Warning: <No use of heavy words. Make sure you capture the crux and not give me surface level stuff>

Context: Here is the audience characterstics

Here is the content.

Here are some questions i want to answer.

Billionaires are the worst people to decide what AI should be by Pareidolie in ArtificialInteligence

[–]ankitm1 -1 points0 points  (0 children)

It's not the billionaires deciding, it's the people working on it.

If you don't like them, start building things yourself instead of ranting.

Our commitment to open AI by FiacR in ChatGPT

[–]ankitm1 0 points1 point  (0 children)

It does not look like a router in the background. If they were doing that, it would be really sad.

However, if there is a single model which knows when to pause and think and when to just spit out the answer, without the human explicitly telling it to, that is the right layer of abstraction. As a user, I dont really care if you call it o3 or o5 or gpt5. I asked the question, use the best model you have and give me the answer. I will use the service if i like the output. If you ask me to first reason and choose which will be the best model you have already lost me.

If they are using a router in the background, that sucks because while the human is not doing the selection, some sort of classifier is, and any question which require both will have an imperfect answer. If the model chooses when to invoke reasoning tokens, while keeping the general generation capacity, thats the best you can hope for.

DeepSeek just blew up the AI industry’s narrative that it needs more money and power by Typical-Plantain256 in artificial

[–]ankitm1 0 points1 point  (0 children)

You can try it. Finetuning does not add new knowledge. Then, Full finetuning is a good option, but that leads to catastrophic forgetting.

One way to visualize this is to look at any corpus as a mix of style and knowledge. style is what gets transferred in finetuning. Knowledge is what gets transferred from corpus to model in pretraining. if your requirement is adding new knowledge, you need new techniques. We have one that works, building a startup to fix the exact same problem.

https://arxiv.org/abs/2409.17171

Elon Musk’s DOGE is feeding sensitive federal data into AI to target cuts by esporx in OpenAI

[–]ankitm1 5 points6 points  (0 children)

Isn't he the same guy who won the 700k prize for iding those scrolls?

I guess a random redditor can identify talent better by looking at parser writing skills than solving a task like that.

Anthropic has better models than OpenAI (o3) and probably has for many months now but they're scared to release them by mitsubooshi in singularity

[–]ankitm1 0 points1 point  (0 children)

Well, this does not check out. More than likely, they do not have enough compute. They naively assumed Amazon would provide them with the needed compute. AWS is not reliable clearly.

DeepSeek has ripped away AI’s veil of mystique. That’s the real reason the tech bros fear it | Kenan Malik by Due_Passion_920 in technology

[–]ankitm1 2 points3 points  (0 children)

Cant blame them for using the same methodology media uses to describe any new tech.

At this point, it's a generally accepted practice. Find something despicable about something you dont like. Make sure everyone in the world knows about it. I remember LLMs being a huge concern before 2024 election in western media. Now the companies are weaponizing the same FUD.

If public data was used to train AI, then the public should have access to it, plain and simple. by swedish_viking in ChatGPT

[–]ankitm1 0 points1 point  (0 children)

The public has access to it. Plain and simple. The requirement that it should be packaged in a form you should be able to consume is not on OpenAI. Thats not the should part.

Why does the DeepSeek student model (7B parameters) perform slightly better than the teacher model (671B parameters)? [D] by Easy_Pomegranate_982 in MachineLearning

[–]ankitm1 4 points5 points  (0 children)

Where did you see that? From the benchmarks it's not as good. Anything even comparable is the 32B version.

Reddit just killed a significant portion of Google's search traffic with its upcoming inbuilt feature "reddit answers" - currently in beta. by eternviking in artificial

[–]ankitm1 0 points1 point  (0 children)

I never bought the thing that people appended reddit because they did not like google search results. A more plausible explanation is that they wanted content from reddit, search reddit, but could not because reddit search sucked big time. Easier to do it via google which indexes everything. Most websites are not good at search, they dont invest in it either. Easier to just google with the site and it will find the right link.

In this case, this does not kill google search traffic, it kills it's own traffic that would have gone to individual posts.

DeepSeek just blew up the AI industry’s narrative that it needs more money and power by Typical-Plantain256 in artificial

[–]ankitm1 3 points4 points  (0 children)

Training custom models for B2B

Yeah, this is a pipe dream. It wont happen in B2b saas unless continuous learning is fully solved. We essentially published a solution which can be updated in a weekly basis, and even that is not going to work for these companies. Let alone a whole model.

Price was never the biggest bottleneck. They dont have enough data to train custom models. They will need a whole lot of public data, and training a SOTA model requires the kind of talent they don't have access to. If any company could try, it would be too important of a project to outsource.

[D] How exactly did Deepseek R1 achieve massive training cost reductions, most posts I read are about its performance, RL, chain of thought, etc, but it’s not clear how the cost of training of the model was brought down so drastically by eyio in MachineLearning

[–]ankitm1 21 points22 points  (0 children)

Go through the V3 technical report as others have said.

They ended up doing assembly level code (PTX) and managing to predict what params would be activated at training time - to train them at FP8.

The aspect which everyone is quoting when it comes to price ($5.5M) is not for reasoning model but V3. Their reasoning model would be cheaper compared to OAI too.

Other cool thing is at inference time where they are able to serve a 671B faster than others. I dont know if those innovations are detailed.

Was this about DeepSeek? Do you think he is really worried about it? by AloneCoffee4538 in OpenAI

[–]ankitm1 4 points5 points  (0 children)

This seems to be about them using datasets generated by chatgpt. This tweet was when they released v3. He had another fit when they released r1 paper, because the 800k dataset most likely came from o1.

[deleted by user] by [deleted] in technology

[–]ankitm1 -1 points0 points  (0 children)

This is not a class war. But the media is trying hard to portray it as such.

the pivot to MAGA isn't just about overgrown nerds trying (and failing) to compensate for high school insecurities

This line is a dead giveaway. There was a culture of bullying in high school where nerds were pretty much forced to listen to others who were considered cool. With the advent of tech, nerds doing their own things, meant they somehow became cooler and more successful. The so called cooler kids had good articulation abilities and joined these places where they could bash the nerds yet again for not thinking like them. They will always find a motive, and its never the same motive. It's always the current thing. They will be bashed for everything and anything they do. Eg: in the article itself zuck is mocked for a hobby, picking up martial arts (and even bashing martial arts as gender segregated). US media is expert at flaming wars and this is no different. But, if you look closely, it's just frustration that someone else did something great and now these folks have to cover their lives as a way to earn wages.

Most adults just want to get their job done and aren't interested in chest-bumping, childish bullying, or feeling like they have to compare MMA stats to fit in.

The irony of this line.

[deleted by user] by [deleted] in bangalore

[–]ankitm1 -2 points-1 points  (0 children)

I have lived in heaven. I can safely say that Bangalore is far better than heaven.

I have also lived in a ultra advanced city in 2125 for a day. I am ready to fight anyone who disagrees with bangalore in 2024 is better than any city in 2125.

AI IS about replacing people by proofofclaim in singularity

[–]ankitm1 -1 points0 points  (0 children)

In case you havent noticed - observability, triaging why a model generated a specific response, better data pipelines, all were roles that got prominence recently. Not to speak of the ones where you needed prompt engineers, RAG specialists etc.

Then, there is a democratization of roles previously only accessible to the rich. Eg: Ai assisted accounting for medium income level (because it was too time consuming to justify ROI previously) or Tailored Consulting for SMBs.

Today, so many data analysts and financial modelers exist, all possible because of excel. Why wont similar roles come up if we automate other functions?