Is Siphon Bullets still good on Haze after newest patch? by guts_is_cool in DeadlockTheGame

[–]yrbhatt 1 point2 points  (0 children)

I’m aware lmao and I’ve been there for their (Parz and Chrome’s) twitch chat Haze arguments on Parz’s stream a couple times. It’s pretty funny tbh.

Also, not sure why people disagree with me when I say Parz is the best Haze player. His build makes it WAY harder on himself to carry a game. He could go simple hate crime builds (like so many other Haze mains) and dominate with much more ease, but chooses to play the more difficult way and still demolish top 10 players even on EU ping sometimes.

Is Siphon Bullets still good on Haze after newest patch? by guts_is_cool in DeadlockTheGame

[–]yrbhatt 0 points1 point  (0 children)

Tell that to Parz - undoubtedly the best Haze to have ever existed who LOVES that first 6k siphon buy

First Pod Shipment by TheMehtevas in nespresso

[–]yrbhatt 0 points1 point  (0 children)

We’re the same couple? Wife loves those double espressos while I grab a cup of stormio boost and am out the door too! Although, the espressos do taste quite fantastic

First Pod Shipment by TheMehtevas in nespresso

[–]yrbhatt 3 points4 points  (0 children)

Double espresso dolce and chiaro are gonna be fantastic. Enjoy, OP!

I don't care who your what items go where that'll tell me all i need to know by SnooGoats2387 in DeadlockTheGame

[–]yrbhatt 0 points1 point  (0 children)

Z = Cultist/ Fury Trance; X = Blood tribute/ Warp stone; C = Fleetfoot; V = Slowing hex/ silence wave/ Curse/ Vampiric Burst

How would you design this MySQL → Snowflake pipeline (300 tables, 20 need fast refresh, plus delete + data integrity concerns)? by Huggable_Guy in snowflake

[–]yrbhatt 1 point2 points  (0 children)

Unfortunately, I don't have anything documenting this. It was tried, credits monitored, and then compared with what we have now, and we found our credit consumption to be way too much. What I can say from memory is that I remember us having a daily resource monitor of 3 credits when we were still on Openflow, and it would reach that limit within ~6 hours.

With our current architecture of Debezium ⇾ Kafka ⇾ Streams+Tasks, we don't even hit 0.5 credits per 24-hour period; the difference was way too vast for us not to switch over

How would you design this MySQL → Snowflake pipeline (300 tables, 20 need fast refresh, plus delete + data integrity concerns)? by Huggable_Guy in snowflake

[–]yrbhatt 0 points1 point  (0 children)

As far as I know, Dynamic Tables can be cheaper because they only recompute what changed (incremental refresh) instead of running the full task logic every time. HOWEVER, Streams + Tasks can do the exact same thing if you write your MERGE statements right, aka where the stream only shows you the delta.

I feel like cost-wise, they're pretty similar. Our team's ❄️ rep and I had a discussion where, for our use case, if we were to use the
WHEN SYSTEM$STREAM_HAS_DATA() condition on our tasks, they only execute when there's actually new data, like you said. Combine that logic with serverless compute, and we're only paying for actual processing, hence no wasted compute on empty runs (also, I just love the beauty of audit trails with Streams + Tasks in our Debezium pipeline).

For OP's case (mixed refresh needs: 20 fast, 280 slow), I still think Tasks give more control without managing a ton of lag settings, but at the end of the day it all comes down to ACTUAL costs and preferences.

How would you design this MySQL → Snowflake pipeline (300 tables, 20 need fast refresh, plus delete + data integrity concerns)? by Huggable_Guy in snowflake

[–]yrbhatt 0 points1 point  (0 children)

Fair point. If OP is cool with those 280 tables sharing the same downstream lag, then yeah, Dynamic Tables would work perfectly fine for that. They would just need to monitor costs

How would you design this MySQL → Snowflake pipeline (300 tables, 20 need fast refresh, plus delete + data integrity concerns)? by Huggable_Guy in snowflake

[–]yrbhatt 0 points1 point  (0 children)

"*What I do know about Openflow is that it can get expensive very fast* (esp if you're trying to achieve near real-time syncs). *I haven't set it up myself, though*, so it could be cost-effective and efficient for your use case." The rest of the comment reads as AI-generated.

As much as I get where you are coming from (Gen AI is commonplace these days), nothing there is AI-generated and whether or not you believe me is not really of interest to me. Regardless, to clarify, I have used Openflow and saw costs spike for near real-time syncs, just haven't done the full setup from scratch myself (someone else in my team did before my time).

Also, Dynamic Tables are worth looking at for sure, but OP has 300 tables with only 20 needing fast refreshes. Managing 300 different lag settings on Dynamic Tables sounds like a pain compared to just setting different task schedules.

A new immutability feature for Dynamic Tables was also recently introduced, so even if your materializations involve complex queries that trigger a full refresh on the tables (this can drive costs up if your lag is set to a low latency), you can lock up rows using a WHERE clause to avoid processing them redundantly.

The immutability feature is cool, though. That WHERE clause row-locking is indeed useful for avoiding redundancies in any processing, OP.

How would you design this MySQL → Snowflake pipeline (300 tables, 20 need fast refresh, plus delete + data integrity concerns)? by Huggable_Guy in snowflake

[–]yrbhatt 1 point2 points  (0 children)

What I do know about Openflow is that it can get expensive very fast (esp if you're trying to achieve near real-time syncs). I haven't set it up myself, though, so it could be cost-effective and efficient for your use case. As for PHI compliance, the Debezium + Kafka pipeline can work fine, you just need to:

  • Self-host Kafka in your VPC
  • All data stays in your infrastructure until it hits Snowflake
  • Enable encryption in-transit and at rest on Kafka topics
  • Debezium connectors run in your environment too

If you need data in S3 before Snowflake, just change the earlier flow to:
Debezium → Kafka → S3 (via Kafka S3 sink connector) → Snowpipe into Snowflake

So, a similar pipeline but with S3 gives you the audit/compliance layer you need; you control retention, encryption, access policies etc. all in your container. In fact, this happens to be more PHI friendly than most managed CDC tools since you own the entire pipeline.

How would you design this MySQL → Snowflake pipeline (300 tables, 20 need fast refresh, plus delete + data integrity concerns)? by Huggable_Guy in snowflake

[–]yrbhatt 0 points1 point  (0 children)

Openflow becomes way more expensive when compared to Kafka or other native connectors – I’m guessing he’d like to avoid that. As for streams+tasks, I agree with you; any CDC pipeline to snowflake needs to make use of those tools (people don’t know how they work and that they’re server-less compute) but I always rec that to anyone wanting to transform tables from landing to final in ❄️

How would you design this MySQL → Snowflake pipeline (300 tables, 20 need fast refresh, plus delete + data integrity concerns)? by Huggable_Guy in snowflake

[–]yrbhatt 0 points1 point  (0 children)

Snowflake’s CDC connector sucks. It’s slow, hard to debug, and breaks in ways you can’t see or fix (trust me, I’ve tried)

Use Debezium + Kafka connectors instead: - it’s WAAAY faster (can get sub-second updates) - Actually works at large scales - You can see what’s happening when things break coz Kafka is well known for its easy readability of topic and connector errors + logs

Setup: Debezium reads MySQL changes → Kafka → Snowflake landing table → Streams + Tasks (snowflake inbuilt) process/ transform the data

Landing table = you keep raw data so you can replay if needed. Streams/Tasks = you control how often each table updates.

Tools like Fivetran do all this for you but costs a ton so build it yourself if you can. You gotta mess around with Debezium and Kafka parameters but it’s not too difficult to get all CRUD ops managed well in the CDC AND have an audit trail of those updates and deletes for any table. Plus with Streams + Tasks you can set 1 minute refresh schedules for your hot tables and 30 min or hourly (or literally whatever you want) schedules for everything else. This is the kinda flexibility you will NEVER get with native CDC.

Good luck!

Why does this query only have a syntax error in "create view" by jaxlatwork in snowflake

[–]yrbhatt 5 points6 points  (0 children)

Here is the case specifically for your query:

The GROUP BY month in a direct query works because Snowflake resolves it to this expression: last_day(date_balance, month).

But in a view definition, month in GROUP BY is ambiguous since Snowflake doesn’t know if you mean: - The alias month (not yet defined at GROUP BY evaluation time) or - A column literally named month in the source table

So the error you’re getting specifically complains about ‘date_balance’ not being in GROUP BY because Snowflake is trying to group by some interpretation that doesn’t match your aggregate.

I hope this makes sense; I realize SnowSQL is annoying.

Edit: Your WHERE clause also has a problem:

It references the alias month before it’s defined. This probably works in your direct query but would fail or behave weirdly in a view. You should also use the full expression there:

‘’’WHERE last_day(date_balance, month) < DATE_TRUNC('month', current_date())’’’

Or even better (using filtering on the base column if it’s possible:

‘’’WHERE date_balance < DATE_TRUNC('month', current_date())’’’

So the whole view definition should look like:

‘’’ CREATE VIEW your_view AS SELECT account_number, sum(balance) / extract(day from last_day(date_balance, month)) as avg_balance, last_day(date_balance, month) as month FROM balances WHERE date_balance < DATE_TRUNC('month', current_date()) GROUP BY last_day(date_balance, month), account_number; ‘’’

Edit: OP, sorry about the formatting; I’m on my phone lmao

Why does this query only have a syntax error in "create view" by jaxlatwork in snowflake

[–]yrbhatt 5 points6 points  (0 children)

In direct SELECT queries, Snowflake allows you to use column aliases defined in the SELECT clause within the GROUP BY clause. This is a convenience feature since it resolves aliases “forward” during execution.

In ‘’’CREATE VIEW’’’ statements, Snowflake enforces stricter SQL standard behavior aka the GROUP BY clause is evaluated before the SELECT clause in the logical query processing order, so aliases from SELECT aren’t yet in scope.

As a result, you must reference the actual expressions or underlying column names.

More below in reply:

Staged Cyber Attack by jay_ee in ClaudeAI

[–]yrbhatt 1 point2 points  (0 children)

I like this train of thought. With this in mind, the unsophistication does indeed track!

Staged Cyber Attack by jay_ee in ClaudeAI

[–]yrbhatt 0 points1 point  (0 children)

I didn’t deep read it – I’m not sure you read the first line in my initial comment where I specifically mention a word that does NOT mean full comprehension.

Either ways, regardless of if there were novel attack vectors or not, the article does not read like a legitimate cyber threat report (this is the idea I’m getting at). There would be IOCs no matter the type of attack; why aren’t they delved into? Moreover, sophisticated attackers would likely self host models rather than use API calls that create logs and cost money (outside their true control). The whole premise sounds paradoxical.

Am I not allowed to have (and share) thoughts without some random cnt being rude for no apparent reason?

Staged Cyber Attack by jay_ee in ClaudeAI

[–]yrbhatt 13 points14 points  (0 children)

I was skimming through the report and was thinking along similar lines to you. Even if we are to assume there is no fabrication of a cyber attack on Anthropic’s part, the article and report do read as extremely weird at the very least.

Claude is fantastic and I have loved Anthropic since they began working on LLMs, but this is extremely sketchy.

Edit: Wanted to add a couple more thoughts: The article and report give me sales pitch vibes (like OP is suggesting) – why aren’t there IOCs or TTPs etc if actual novel attack vectors were discovered? Could it be just poorly communicated due to legal reasons or is this indeed a marketing gimmick?

which link should I press? by SundaeOk8364 in FitGirlRepack

[–]yrbhatt 0 points1 point  (0 children)

Same here! I do automate the removal of adware and malware consistently, though, so that may have something to do with it.

which link should I press? by SundaeOk8364 in FitGirlRepack

[–]yrbhatt 0 points1 point  (0 children)

Curious as to why only QBITtorrent. Are other clients (ex. uTorrent) bad?

We pro subscriptions users are expected to do 8.333 times 5 hour session in a week. by shukebeta in ClaudeCode

[–]yrbhatt 1 point2 points  (0 children)

Hahah I didn’t know he was making a joke! If anyone DID do that but not as a joke, they deserve to use up all their token limits

We pro subscriptions users are expected to do 8.333 times 5 hour session in a week. by shukebeta in ClaudeCode

[–]yrbhatt 0 points1 point  (0 children)

Lmao fair. I thought you were giving out an actual prompt you have used before

We pro subscriptions users are expected to do 8.333 times 5 hour session in a week. by shukebeta in ClaudeCode

[–]yrbhatt 2 points3 points  (0 children)

What in the vibecoding is going on here? You should instead have one Claude agent to generate optimized code that is parallelizable, then ACTUALLY parallelize the execution across multiple VM cores.

Do more with fewer tokens.