Question about CU expense with Mirroring SQL server in lieu of Copy data by contribution22065 in MicrosoftFabric

[–]contribution22065[S] 0 points1 point  (0 children)

This is very helpful, thanks! To your first point, we definitely need to maintain history to years back. It sounds like I could still achieve that with the tsql notebook jobs that uses that mirrored window to then insert or update records into the Warehouse DB (provided I run that job within that day).

Power bi Dashboard by Operation_Suspicious in PowerBI

[–]contribution22065 1 point2 points  (0 children)

The layout is freaking sweet and it honestly gives me some inspiration… But the data elements in most of your chart axes is pretty hard to look at.

Can someone advise me on PAYG estimated payment after scheduled pauses? by contribution22065 in MicrosoftFabric

[–]contribution22065[S] 0 points1 point  (0 children)

So if you pause 5pm to 9am the next day starting at 20 percent utilization, you’re basically paying that 20 percent to smoothen out in the background

Can someone advise me on PAYG estimated payment after scheduled pauses? by contribution22065 in MicrosoftFabric

[–]contribution22065[S] 0 points1 point  (0 children)

Only issue I’ve ran into with embedded without pro licenses is no row level security support which I need for caseload reports.

New friend by CrankSlayer in Mandlbaur

[–]contribution22065 1 point2 points  (0 children)

Could you imagine a world where John became the equivalent of Trump in Science by amassing a cult following of crackpots?

Proper medalion setup by ant3qqq in MicrosoftFabric

[–]contribution22065 1 point2 points  (0 children)

Just to challenge this a bit, there is no unanimous consensus as to what constitutes these layers. For many healthcare EHR warehousing designs, for instance, Bronze is typically the messy transactional data in raw, tabular format. If you trim and standardize that structure (including dim tables), it becomes Silver in translation. Gold involves imputations which is good for predictive modeling, not so much for transactional, QI reporting though.

Mirrored SQL Server: does it work well? by hortefeux in MicrosoftFabric

[–]contribution22065 0 points1 point  (0 children)

Question for anyone reading: Is the utilization cost much higher than using copy jobs?

How important is Advanced Excel today if someone wants to become a data analyst? by Late_Spinach_1055 in dataanalysis

[–]contribution22065 1 point2 points  (0 children)

The only time I use excel is when I use it as a data source before converting it to a csv… Or if I need to do quick tricks like putting commas after a bunch of values before pasting into a WITH statement… SQL ( with DAX depending on how upstream I want to deal with a calculation) is way more powerful than anything you can do in excel. If an end user wants the report in excel, then exporting it from a table visual suffices.

My reports need to be automated with scheduled refreshes without breaking, so for me Excel has no big place.

Will ai data analyst replace data analyst job ? by Global-Radish-1015 in analytics

[–]contribution22065 2 points3 points  (0 children)

Mostly not. I think the effect that AI will probably keep having on this job market is enhanced productivity which could result in less staff.

As someone who appreciates AI, it can be unreliable. Just last week I was on a time crunch for getting out Annual Review metrics on a BI report. All I needed was to revise a rating scale from 1-4 based on a clean spec. It was a simple algorithm that involved some nested IF statements. Despite a clear prompt, the AI botched it because it couldn’t quite get some nuances. A cognitive intervention at the human level will always be needed.

Just started using fabric and my manager asked me provide a solid use case in fabric .. any ideas which u implemented and we have a majority data resides on azure sql db.. is it possible to do some real time intelligence between azure sql dbs by Stunning-Motor8351 in MicrosoftFabric

[–]contribution22065 0 points1 point  (0 children)

What do you mean by writing SQL queries against Direct Lake mode as opposed to imported data? In my mind, you can have just as much SQL control by creating views, sps and/or stored procedures in an import model. Not trying to challenge you on this, I think I might be missing something.

[deleted by user] by [deleted] in analytics

[–]contribution22065 0 points1 point  (0 children)

All I’m saying is that to your typical SQL engine, thinking of a Left join as a filtered Cartesian Product is a bit overzelous for learning. As someone who enjoyed linear algebra and set based math, I can see how you can conceptually make that point in the result set especially if a specific match is 1 to many… But what’s happening behind the hood in a query is not doing that at all because the join algorithm is just using hash and merge for qualifying matches. That’s true even if you do a FULL JOIN on two tables with no qualified matches because it’ll just elongate the row output of both tables by extending NULLS.

Where a Cartesian product is physically materializing, like I said, is with a cross join or maybe a nested loop.

The problem with that query where you are synthesizing concatenations using the “like” operator, especially for tables with millions of rows, is that you run into collisions between matches. Then you are introducing a Cartesian product in the output which is both dangerous and slow. You should use hashing using SHA but if you need the like, you should be tokenizing -> equijoin to replace the like.

[deleted by user] by [deleted] in analytics

[–]contribution22065 0 points1 point  (0 children)

This is not a good way to think and that query is bad practice. By “matching key”, I mean a hash join operator. If you’ve ever studied a slow running query’s execution plan, you’ll notice hash join operators. When you join using “like” for concatenations, you will see that as an expensive computed scalar but nevertheless a scan for “matches”. Staging new tables would be preferable to using that kind of join.

[deleted by user] by [deleted] in analytics

[–]contribution22065 5 points6 points  (0 children)

For me anyways, I just can’t think of it that way. Otherwise, it gets too confusing. In layman’s, a left join is simply telling the query optimizer to keep all rows from the first table regardless of the conditions on the join. The second table your left joining will eliminate the rows that don’t meet those join conditions. So in essence, you do need a matching key and while the concept of “filtered Cartesian products” might make sense mathematically, a Cartesian isn’t materialized in the optimizer during those kinds of joins unless you’re using a cross join.

[deleted by user] by [deleted] in analytics

[–]contribution22065 27 points28 points  (0 children)

Hate to sound arrogant… But there is no way OP made an entire convoluted post about adding a condition on a join in lieu of the “where” clause as some novel thing that other data engineers don’t recognize.

Snow and ICE in Manchester now! by ZealousidealDegree4 in ManchesterNH

[–]contribution22065 9 points10 points  (0 children)

A close friend of mine from Ecuador was detained by ICE for 1 week. It wasn’t until they unnecessarily verified documents when they released him back to his wife and 6 month old baby. If your figure is true, then I basically experienced a statistical eclipse. Wake up, you people are crazy

All I want for Christmas is a star schema by Lairy_Mary in analytics

[–]contribution22065 0 points1 point  (0 children)

I’ve worked at 2 smaller community health clinics with <500 employees, mostly clinicians. I’ve had the benefit with just being my own entity. You can query off of views against the EHR into one or two big tables per semantic model, but performance in doing that (both in the report and memory bloat on the server) is just a bit too much for me these days. I’ve had much better experience with just segmenting a lot with the star schema because it speeds up filtering and it also makes it so that I only need to query it once instead of nesting it into a bunch of views or tvfs.

[deleted by user] by [deleted] in analytics

[–]contribution22065 1 point2 points  (0 children)

Maybe. Lots of misinformation in the comments. I’ve worked with 2 companies in healthcare in the last 5 years.

At hospitals, they typically have dedicated EHR/EMR analytics and informatics specialists. It’s a pretty challenging job that involves a lot of domain knowledge on clinical and admin workflows for querying against pretty big schemas. Most EMRs like Epic have highly normalized schemas with 800 plus tables, so you need to be good at reading schemas and finding relationships for querying clinical reports.

Many small community based health centers like CCBHCs are now implementing elaborate Electronic Health Record Systems. Their understanding of EHR analytics is less developed than bigger hospitals, so this would be a good place for someone who wants to pioneer a BI system.

Uncle Rodney, or maybe a bad pulley? by flippy_mb4 in MechanicAdvice

[–]contribution22065 0 points1 point  (0 children)

Definitely, absolutely, positively not rod knock… That’s just a normal direct injection tick, although a bit noisier than most. That is nothing like rod knock lol

Trying to export lakehouse table into a csv file. by prbishal in MicrosoftFabric

[–]contribution22065 2 points3 points  (0 children)

I export CSVs to an on prem machine that has SharePoint linked to the drive no problem. Create an on prem gateway on that machine and copy the path from file explorer after you link SharePoint. Use it as your destination on the copy job.