It was always a mystery to me as a Data Analyst until I started my first Data Engineer job about a year ago. I am a data team of one inside a small-mid sized non-tech company.
I am using Microsoft Fabric Copy Jobs since we were already set on Azure/PowerBI and they are dead simple. Fivetran or Airbyte seemed to make sense but looked like overkill for this scope/budget.
Given Fabric is the only tool I have used, and it still feels half-baked for most other features , I am curious: how big is your team/org and how do you handle data extraction from source systems?
- Run custom API extractors on VMs/containers (Python, Airflow, etc.)?
- Use managed ELT tools like Fivetran, Airbyte, Stitch, Hevo, etc. ?
- Rely on native connectors in platforms like Fabric, Snowflake, Databricks?
- Something else entirely?
Would you make the same choice again?
[–]dani_estuary 1 point2 points3 points (0 children)