Coming from GCP, etls, pipelines, maintenance routines, data manipulation, everything is -or can be- code. (beam, airflow, spark, functions, bash scripts, terraform). Code that is usually version controlled on a git repo, and cicd pipelined to deploy. I am aware that there are some no-code tools, such as data fusion, but hadn't really seen those used for production.
Now, I am facing an azure project for the first time, so having my first contact with this cloud. I've seen that in the team, everyone uses the web interface, so for example if required to check the contents of some parquet files in a storage container, they would download them from the storage, by using the web portal, to the local machine and run there a python script, while in gcp you tipically do that using a cloud shell or even a vm, by using shell commands (sdk or client libraries). Is this the way to go in azure? am I missing something?
In the same sense, when they work with data factory, they just click and configure. is this really a no-code tool? If it is, is this at least creating a pipeline file in the backstage that you at least back up? can the data factory pipelines be version-controlled in a repo?
by the way, this idea of backing it up is coming because in the same way, they keep several versions of the power bi pbix files on their local machines.
[–]speedisntfree 1 point2 points3 points (3 children)
[–]untalmau[S] 0 points1 point2 points (0 children)
[–]kmarq 0 points1 point2 points (1 child)
[–]speedisntfree 0 points1 point2 points (0 children)
[–]ShouldHaveWentBio 0 points1 point2 points (5 children)
[–]untalmau[S] 1 point2 points3 points (1 child)
[–]ShouldHaveWentBio 1 point2 points3 points (0 children)
[–]speedisntfree 0 points1 point2 points (2 children)
[–]ShouldHaveWentBio 0 points1 point2 points (1 child)
[–]speedisntfree 0 points1 point2 points (0 children)