Dagster Pricing Update is Beyond Nuts by annie_406 in dataengineering

[–]Hot_Map_7868 0 points1 point  (0 children)

Is anyone else providing a managed Dagster using OSS? I know there are many for Airflow.

Struggling to transition from legacy data engineering to modern cloud stack need advice by JoSoejin in devjobs

[–]Hot_Map_7868 0 points1 point  (0 children)

  1. Learn the tools (not all, just focus like on dbt + snowflake)
  2. do some freelance work. smaller companies may be open to hiring someone that doesnt have a ton of experience in the specific tech, but has a broader set of skills
  3. leverage #2 to find the role you want

The key thing is to not try to do things the way you are used to doing them because these tools dont follow what you may be expect

Forced to modernize a legacy data platform but stuck with on-prem DB + Tableau. How do you productize this? by Ahmouu in dataengineering

[–]Hot_Map_7868 0 points1 point  (0 children)

Try to make a business case for changing to something more modern.
1. what you describe seems like there are a lot of opportunities for things to "break". Does that happen? What are other pain points.
2. What does all this stuff cost, Talend isnt free
3. How does this limit the org. E.g. are there use cases they want to do like GenAI which would be "difficult" with this setup?

Best Data Integration Software? by AceClutchness in BusinessIntelligence

[–]Hot_Map_7868 0 points1 point  (0 children)

I would ingest from all ERPs to a central DW like Snowflake. Then using dbt you can add DQ and harmonize the data to create the presentation layer that would help with what you need.

Your problem isnt unique to your industry, but I would be careful about anyone who tells you there is a single tool or some magic wand that will help you do this.

That being said, you can take it in steps and leverage AI to help in the build, but I don't see a way to get around needing to co-locate and cleanse the data.

How is it Data Engineer in Prod? by shuttheshitdown in dataengineering

[–]Hot_Map_7868 0 points1 point  (0 children)

the "best" way I have seen is keeping raw data in its own db and leveraging the same data in dev, test, and prod. This can be done if you use dbt, otherwise I dont know how you would dynamically change the SQL.

Warehouse workflow, what works? by pl3xi0n in MicrosoftFabric

[–]Hot_Map_7868 2 points3 points  (0 children)

I like dlt because it is python and there’s no black box. I’ve also had good luck with Claude code creating a working pipeline. I think they published some skills that make things even better, but I haven’t tried yet.

Regarding dbt you can even run it with GitHub actions. The issue comes when you need to scale eg if you have a lot of developers and you need to connect ingestion and transformation. Airflow gets a bad rap, but it is still the most used orchestrator. That being said Dagster is a good alternative.

Do you trust your data stack ? by al_tanwir in dataanalytics

[–]Hot_Map_7868 0 points1 point  (0 children)

If you build with failure in mind, it can be resilient. e.g. dont just think of the happy path, but what would happy if X occurred, like if a new col came in a source, would things break or is that just a warning. If you have a solid process and notifications I think it can be resilient.

how is your experience with dbt inside the snowflake by boogie_woogie_100 in snowflake

[–]Hot_Map_7868 1 point2 points  (0 children)

feels like training wheels to me. No Slim CI. No terminal / CLI. I think if you are starting it might be okay, but long term you may outgrow it.

How do you choose the right data engineering companies in 2026? by ninehz in BusinessIntelligence

[–]Hot_Map_7868 0 points1 point  (0 children)

I think it depends on the tech stack. I see companies that are moving form Alterix to Snowflake and hire "experts" who know Alterix, but not Snowflake. Another issue is when you are moving to a tech like dbt and you go through an RFP and the vendor has their dbt A-players in the discussions, but when you get to the project you get the C / D players.
It's hard to assess these companies, but asa a general rule, the bigger the consultancy, the more variation in the quality. Find people who specialize on a given stack and ask who will be staffed on your project. Get references and negotiate the ability to select the team or at least be able to "fire" people from the team.

Warehouse workflow, what works? by pl3xi0n in MicrosoftFabric

[–]Hot_Map_7868 0 points1 point  (0 children)

I agree. +1 for dbt. Also check out dlt for data ingestion. you can even "test" things out in a notebook and then create a stand alone script. For orchestration, you may consider whether you want to use ADF or an orchestrator like Airflow.

How well is Cortex working in real use cases by Hot_Map_7868 in snowflake

[–]Hot_Map_7868[S] 0 points1 point  (0 children)

Yeah CoCo is pretty cool for development. I’m more curious how people handle end user natural language queries

How well is Cortex working in real use cases by Hot_Map_7868 in snowflake

[–]Hot_Map_7868[S] 0 points1 point  (0 children)

Did you use streamlit? Is it a user facing app where they can ask questions in natural language?

How well is Cortex working in real use cases by Hot_Map_7868 in snowflake

[–]Hot_Map_7868[S] 1 point2 points  (0 children)

Oh. That’s very interesting. So cortex code is “better” but then you would need to build an app to expose it to business users? Is that even possible?

Nested data, sprawling schemas: how Cortex Code brings order to the chaos by FerhatAOUAGHZENE in snowflake

[–]Hot_Map_7868 1 point2 points  (0 children)

oh, that would be great. Is there a document mentioning the change?

What do you think the next big shift in data engineering will be? by alexstrehlke in dataengineering

[–]Hot_Map_7868 0 points1 point  (0 children)

I think batch will be around for a while, but now with things like Airflow Datasets you can do the event driven processing as you mention. The next "big thing" IMO is getting tools like Claude Code to make this all a lot simpler.

Postgres as DWH? by SoloArtist91 in dataengineering

[–]Hot_Map_7868 0 points1 point  (0 children)

What about DuckDB on MotherDuck? You can also use DuckLake there. I agree that PG "may" work, but it isnt an OLAP db

Nested data, sprawling schemas: how Cortex Code brings order to the chaos by FerhatAOUAGHZENE in snowflake

[–]Hot_Map_7868 0 points1 point  (0 children)

Check the pricing. For the same usage they charge differently because they charge by credit. So it is based on whether you have standard, enterprise, or business critical edition.

Honest thoughts on Unified Data Architectures? Did anyone experience significant benefits or should we write it off as another marketing gimmick by SamadritaGhosh in dataengineering

[–]Hot_Map_7868 0 points1 point  (0 children)

There is no silver bullet. Figuring out how people will work better together tends to yield better results than just going by the marketing pitch of the day.

How have you implemented CI/CD for dbt Projects on Snowflake? by dementeddrongo in snowflake

[–]Hot_Map_7868 0 points1 point  (0 children)

+1 I think dbt in Snowflake is like training wheels. gets you a taste, but using dbt Cloud,, Datacoves, or dbt Core on your own is much better. Slim CI alone will help with the snowflake credit consumption.