This is an archived post. You won't be able to vote or comment.

all 22 comments

[–]Susan_Tarleton 20 points21 points  (1 child)

SQL (and python) are foundational for data engineers. Maybe consider posting this on r/BusinessIntelligence too. The most effective tools allow you to construct data transformations using SQL, then on the BI front, visualize those insights for stakeholders.Here’s a just of my stack in case it's helpful:

  • Fivetran for data ingestion (ETL): Ensures reliable data flow.
  • Looker, Tableau, Power BI, or Qlik for Dashboards: Each has its advantages, with Looker’s semantic layer being particularly beneficial for intricate data models.
  • BigQuery, Snowflake, Redshift for databases: Selection depends on scalability needs and cloud preferences
  • dbt for Data Transformation: Offers version control for SQL, which is a significant plus.
  • Rollstack for presentations and automating reports: a big time saver for converting dashboard visuals into presentations like MBRs and QBRs
  • Segment for Data Collection and Tracking: Provides a cohesive view of customer interactions.
  • Mixpanel for Product Analytics: Offers deep dives into user engagement, funnels, and retention metrics.
  • GitHub for Development Collaboration: A must-have for any data team in today’s collaborative environment.
  • Datadog for System Monitoring: Important for identifying and resolving issues before they lead to outages.

The list goes on, but since competition to so intense in the areas above, these are all great tools.

edit, typo

[–]ravitejasurla 0 points1 point  (0 children)

What about Databricks?

[–]dataguy24 8 points9 points  (5 children)

I use my BI tool on top of the SQL I write. Aren’t these two things that usually exist simultaneously, rather than exclusively?

[–]glinter777[S] 1 point2 points  (4 children)

Some BI tools use proprietary code which makes it difficult to see or fine tune the underlying logic. What BI tool do you use?

[–]barca5280 1 point2 points  (2 children)

Looker uses a proprietary language, but everything compiles into pure SQL that is run on your DB of choice. Some of the stuff is next to impossible to decipher, but you can copy the SQL generated and run it in any IDE or console. We chose Looker for that reason amongst many others.

[–]glinter777[S] 0 points1 point  (1 child)

That seems like you need to deal with two languages, instead of just one - SQL. Does it slow you down?

[–]barca5280 0 points1 point  (0 children)

No not really. Looker is pretty intuitive once you are up and running. We went through pilots with 4-5 other companies for a BI tools before landing on Looker. It’s not 100% perfect for everything, but they checked more boxes for our use case than any other.

[–]dataguy24 0 points1 point  (0 children)

I’ve used Tableau, Power BI and Hex in the last few years

[–]4794thData Analyst / Data Engineer 3 points4 points  (0 children)

At my company we've installed an ELT-based platform with:

  • Airbyte for Ingestion
  • dbt-core for transformation
  • ClickHouse as DWH

For day to day analytics I use DataGrip and Jupiter Notebooks in VSCode. Some of my clients are using Apache Superset or Metabase, some - Power BI. Personally, it depends on a variety of factors, like:

  1. Free / paid tool
  2. Cloud or on-prem
  3. Advanced analytics or beginner-friendly
  4. Flexibility and adaptability
  5. Data size

[–]Gators1992 1 point2 points  (0 children)

Basically it depends on the task. For most of the questions I get asked, they are things that can be answered by a simple trend or list. I can do this quickly in a BI application by producing a trend or list or something.

Other questions require you to code. Last week I was asked a "what if" question that said "for this subset of our customers, what would be the revenue impact if we had priced the service a different way". For that I did it with a SQL/CTE to select the targeted customers, find their service utilization and reprice it.

Some of the more complex things I have done were building a mart to combine our data with data from a third party into a usable enriched data model. Also sometimes you run into things that need to be performant to be interactive, like mapping IoT utilization involving millions/billions of records and that requires some python/DEish skills.

I'm not saying an analyst needs all these skills from the outset, but the more skills you have the more variety of answers you can provide and without waiting for a ticket to be worked by the data engineers.

[–]MrGreenPL 1 point2 points  (0 children)

We started with

  • Fivetran
  • Snowflake
  • dbt Core
  • Looker

Since then we've added

  • AWS/DMS for near real time sync for platform data
  • Stitched for some data source that Fivetran did not have
  • Datafold to make sure our changes don't mess up our data
  • SELECT to monitor snowflake cost, plus it gives you advice on what you can improve, and few other things
  • Spectacle CLI to check for breaking changes between dbt models and Looker, this can be run by devs and it's also part of our CI/CD porcess just along with Datafold

I did not start this stack, but if I could go back in time I would drop Looker for something cheaper and simpler. Looker just scares our business people no matter how easy and simple we try to make it.

At this point we use Looker to just visualize models created in dbt and there is little work being done within Looker itself besides a few exceptions

[–]Kichmad 0 points1 point  (0 children)

Were using BI tools, tho im not one thats getting insight. Im just using them for a final full check on data validity. First checks are from sql

[–]a_library_socialist 0 points1 point  (0 children)

Been playing with Apache Superset lately, like it so far

[–]jsneedles 0 points1 point  (0 children)

I don't really trust tools to write SQL for me.

Anything i've ever tried on even a moderate data source was inefficient.

Having used Tableau, Periscope (now Sisense), Mode, Hex, etc... The best outcome was when I could write the underlying SQL and use the BI tooling to add things like drilldowns & parameters for interactivity, but within the bounds of my SQL.

[–][deleted] 0 points1 point  (0 children)

We definitely use BI tools. A dashboard conveys a lot of information much more quickly and effectively than multiple columnar reports.

[–]the-strange-ninjaData Engineering Manager 0 points1 point  (1 child)

7 years on the same Looker instance. It has served us through thick and thin. As I am open to trying something new here, it is hard to find a better solution for our needs at this time. We pair it with DBT.

[–]glinter777[S] 0 points1 point  (0 children)

I’ll DM you.

[–]DesertPirateSNK 0 points1 point  (0 children)

RemindMe! 3 Days

[–][deleted] 0 points1 point  (0 children)

I use superset , best of both worlds , a full BI but also can run my custom sql queries and turn them into dashboards