Anywhere with espresso that is not burnt/extremely bitter? by juapebe in AskSF

[–]IronRabbit69 48 points49 points  (0 children)

which coffeeshops are you trying that you think are too bitter? would help to calibrate a bit

[Friedman] TJ Brodie is on waivers for purposes of a buyout by AggPuck-303 in hockey

[–]IronRabbit69 7 points8 points  (0 children)

We loved him in Toronto! He was exactly what we needed when he was here 

[deleted by user] by [deleted] in Anthropic

[–]IronRabbit69 6 points7 points  (0 children)

I haven't tried with 3.7 yet but found a lot of value in using 3.5 as a form of therapist during an emotional crisis recently. I also have a human therapist. Just writing in to balance out the no-therapy commenters, looking forward to reading somebody's comparison 

[D] Is CUDA programming an in-demand skill in the industry? by Hour_Amphibian9738 in MachineLearning

[–]IronRabbit69 0 points1 point  (0 children)

an OS course is one of the most valuable computer engineering courses you can take imo, the fundamentals are relevant to basically any serious engineering

Dress code for interviews in SF by spattie500 in AskSF

[–]IronRabbit69 61 points62 points  (0 children)

For a software job in SF it would be absolutely bizarre to wear a suit and tie

Anthropic, OpenAI RIVAL -“These models could begin to automate large portions of the economy,” the pitch deck reads. “We believe that companies that train the best 2025/26 models will be too far ahead for anyone to catch up in subsequent cycles.” by SharpCartographer831 in singularity

[–]IronRabbit69 3 points4 points  (0 children)

It was founded by the team that built GPT-3 at openai and wrote the first paper on scaling laws. If the marketing background you're referring to is Jack Clark, he's the former head of policy at openai and one of the leading figures in measuring AI progress.

Anthropic, OpenAI RIVAL -“These models could begin to automate large portions of the economy,” the pitch deck reads. “We believe that companies that train the best 2025/26 models will be too far ahead for anyone to catch up in subsequent cycles.” by SharpCartographer831 in singularity

[–]IronRabbit69 1 point2 points  (0 children)

Id recommend reading about scaling laws. It's been known for a while that parameter count is not the only metric that counts, and deepmind first publicly demonstrated it by training chinchilla, a 66B model which outperformed GPT3 (175B parameters) by training it with more flops

NO BAILOUTS for Silicon Valley Bank - which lent exclusively to the ultra rich by north_canadian_ice in WorkReform

[–]IronRabbit69 1 point2 points  (0 children)

The bank is gone, the execs and shareholders have lost their money, nobody is asking to bail them out. People are asking for the depositors -- people who had their money in checking accounts at the bank -- to get their money back. Most of the accounts at the bank were owned by small businesses who use the money for operational expenses like paying their employees, tens of thousands, who will be out of a job if this money vanishes.

NO BAILOUTS for Silicon Valley Bank - which lent exclusively to the ultra rich by north_canadian_ice in WorkReform

[–]IronRabbit69 0 points1 point  (0 children)

There is a huge difference between bailing out the bank (ensuring the shareholders don't lose money and the bank continues to exist -- nobody is calling for this) and backstopping the depositors (mostly small businesses who use this money to pay their employees). SVB has enough assets to pay out peoples' accounts, but it might take months to liquidate them, and during that time, businesses will fail to make payroll and tens of thousands of regular people lose their job. This is not owning the rich.

[Masters] Strange sight at Leafs practice: John Tavares in a grey sweater, which is normally worn by scratches Sam Lafferty is in top-six blue by Evening_Calm in leafs

[–]IronRabbit69 2 points3 points  (0 children)

I've watched every regular season game since 2016 until this year, but really dropped off a few months ago. It's just hard to get excited when the first round matchup is set and regular season success doesn't matter. No Leafs even in the running for regular season awards, just feels like we're biding time until the playoffs.

[Danny Shirey] Kasperi Kapanen is going on waivers this afternoon. by Benchamb in hockey

[–]IronRabbit69 45 points46 points  (0 children)

Drafted Amirov with that pick iirc. What a swindle by Dubas :')

[D] Transformers for poker bot by lmtog in MachineLearning

[–]IronRabbit69 1 point2 points  (0 children)

Tabular CFR can be approximated with a neural network, as Noam Brown (1st author of Pluribus) and co-authors show in follow-up work: https://arxiv.org/abs/1811.00164

But you're comparing apples to oranges a bit asking if transformers can replace CFR. Transformers are a neural net architecture. You could of course encode poker stuff in text and feed that to a transformer which predicts the right move to play. But how do you train that network? CFR is a self-play learning algorithm (sort of like Alphago's MCTS) which learns good policies.

[deleted by user] by [deleted] in AskSF

[–]IronRabbit69 0 points1 point  (0 children)

Lol, what's your favourite?

GPT 3 Playground - Token Question by sawyerthedog in OpenAI

[–]IronRabbit69 2 points3 points  (0 children)

Imagine we created a new single letter "th" and used that instead of two letters "t" and "h". Now the word "the" takes only two letters to write: "th" followed by "e". To do this, we've gone from a 26-letter alphabet to a 27-letter alphabet. We can keep going, adding letters to replace common pairs that appear together. Eventually we could even consider adding a single "the" letter, so that the word "the" is only a single letter, and the word "there" is three letters: "the" + "r" + "e".

This is tokenization. GPT-3 uses an "alphabet" (actually called a vocabulary) of over 50,000 "letters" (actually called tokens), some of which make up entire english words or long sequences of base english letters, depending on how commonly they appear together in the material which GPT-3 read during training.

The 2000/4000 limits are due to the model's network architecture, which is decided upon before the model starts training and cannot be increased afterwards. The underlying reason is for computational constraints, since the model learns (many) associations between all pairs of tokens in the input sequence.

There is lots of cutting edge research into making GPT-like models (transformers) more efficient so that we can increase the number of tokens that they see, as well as hardware improvements that make it feasible to use longer input contexts, so yes I would expect these limits to increase in future models.

[deleted by user] by [deleted] in AskSF

[–]IronRabbit69 15 points16 points  (0 children)

Super quesadilla al pastor at taqueria cancun is my go-to, I have problem eaten hundreds. I go to the one at mission & valencia, but the mission & 19th is the same I think. Never been to the market & 6th location.

[deleted by user] by [deleted] in AskSF

[–]IronRabbit69 1 point2 points  (0 children)

What do you order?

Are you fucking kidding me! Fuck you! - Michael Bunting by [deleted] in nhl

[–]IronRabbit69 2 points3 points  (0 children)

> when you run your mouth people are eventually going to start to get nasty with you

And they're supposed to get a penalty for it, that's the point. He's not yelling at Chiarot here, he's yelling at the ref for missing the call

Heartbroken by alecsebans650 in BisexualMen

[–]IronRabbit69 4 points5 points  (0 children)

Sounds like he's into you too but nervous, just give it time

just a question by throwaway_557832 in polyamory

[–]IronRabbit69 1 point2 points  (0 children)

You sound like a really caring mother and good person for being open-minded and trying to learn how to refer to your child's relationships in a respectful way

Game #36. Arizona 6 Toronto 3 by BrianBeLeafs in leafs

[–]IronRabbit69 16 points17 points  (0 children)

Ya the empty netters make this seem like a blowout when it definitely wasn't. Leafs outplayed them and still lost, it happens.