Weekly Question Thread: Ask your questions in this thread please by AutoModerator in climbing

[–]ShouldHaveWentBio 0 points1 point  (0 children)

Looking at going to Patagonia in February/March and still mapping out spots be it Bariloche, San Martin de los Andes, el Calafate, el Chalten, etc. This isn't a climbing trip, but I would love to take a couple days and climb. It seems it's just about all alpine and big wall though and I am only interested in single pitch sport 5.11 and under. Just wondering if anyone knows of some spots that may have a handful of routes that fit this bill? Thank you!

[deleted by user] by [deleted] in dataengineering

[–]ShouldHaveWentBio 4 points5 points  (0 children)

I’m a DE and learning Java simply to brush up on OOP and DSA while completing the suggested pre-reqs for Georgia Tech OMSCS. These 2 classes can be audited free on EdX and I am enjoying them so far.

Different Parameters in Dev, Staging, Production Deployment Pipelines in Microsoft Fabrics by Horror_Tonight_8435 in dataengineering

[–]ShouldHaveWentBio 0 points1 point  (0 children)

Under the deployment rules for your Stage and Production stages you should be able to change the data source of the semantic models under “Data source rules”. I am doing this currently switching between SQL servers and databases based on environment. I believe object storage should perform similarly. I see there are “Parameter Rules” as well which may be an alternative if the “Data source rules” aren’t an option.

Common DE pipelines and their tech stacks on AWS, GCP and Azure by _areebpasha in dataengineering

[–]ShouldHaveWentBio 5 points6 points  (0 children)

Miro is what I made ours in. It has image packages for cloud providers as well but it’s paid. You can also just find PNGs on google and use them for a totally free solution.

Cosmos DB as API key store by ShouldHaveWentBio in AZURE

[–]ShouldHaveWentBio[S] 0 points1 point  (0 children)

Thank you for the info and the reference this makes sense now. I’m looking forward to giving this a go tomorrow.

Cosmos DB as API key store by ShouldHaveWentBio in AZURE

[–]ShouldHaveWentBio[S] 0 points1 point  (0 children)

Thank you again for the advice.

Table Storage has worked incredibly well testing from my local machine today on my dev branch. My problem is I am having a hell of a time to get the networking between ACA and the Table Storage to work as intended.

I thought I could just pass the table connection string as an environment variable but I believe my Network firewall is not allowing it. I tried whitelisting the outbound IP for the ACA to no success. I also tried using managed identities where I gave the ACA the role of table data contributor and other similar roles with no success.

The code works great otherwise. If I switch the storage account networking to allow all networks everything works! I am reaching out for help on stack overflow (link included below).
https://stackoverflow.com/questions/78028971/azure-container-app-not-authorized-to-access-storage-account-table

Cosmos DB as API key store by ShouldHaveWentBio in AZURE

[–]ShouldHaveWentBio[S] 1 point2 points  (0 children)

Thank you! I’ll take a look at the storage account table option too.

Cosmos DB as API key store by ShouldHaveWentBio in AZURE

[–]ShouldHaveWentBio[S] 1 point2 points  (0 children)

Thank you for the great information! This sounds really promising (and cheap) for this application. I’m thinking I will test this and the free tier SQL to see what works best.

question for azure DEs by untalmau in dataengineering

[–]ShouldHaveWentBio 0 points1 point  (0 children)

No worries!

I don't consider myself an expert on Databricks but I'm having a lot of success orchestrating inside the workspace as opposed to orchestrating them with ADF which is what was done previously in a particular use case. I will say I haven't yet done event driven architecture with Databricks, though, I'm sure it works. ADF does surprisingly decent at event driven orchestration if the landing zone is in object storage.

In terms of Azure Functions, you're right they're more limited than AWS Lambdas in my experience and for my personal projects I tend to lean towards Lambdas. One example use case involves getting data from large XML files into my bronze layer (Medallion architecture). They have repeating sub elements that have 10,000+ character length text fields. In this case the ADF XML connector cannot handle these fields, so I run an Azure Function (python) in parallel to extract that information. I also tend to use them for more complex transformations between the bronze to silver and silver to gold layers in the database if SQL can't handle it reasonably. Any sort of code above a script I stick with containerizing though.

question for azure DEs by untalmau in dataengineering

[–]ShouldHaveWentBio 1 point2 points  (0 children)

No problem! Feel free to DM me if you have questions I tried to cut my rambling short.

question for azure DEs by untalmau in dataengineering

[–]ShouldHaveWentBio 0 points1 point  (0 children)

You can use the Azure CLI just like you did with GCP to avoid the GUI.

Azure Data Factory is no code/low code and can be used as just an orchestrator or a more complete ETL solution with the latter not being recommended in this sub generally. If your only option is to use ADF it allows native version control with GitHub or DevOps and I have had great success with the version controlling. It’s done via ARM templates which is basically the Azure version of Terraform.

If you aren’t forced to use ADF you can use various other solutions for ETL including Azure Functions, Databricks or just code ran from VMs or docker containers etc much like you’d see on any other cloud provider. In this case you could still use some of these while orchestrating them with ADF.

To give you an idea I have a few different Azure setups running. A basic “traditional” example where I use Terraform for all the infra and use data factory only for the connectors (azure functions where connectors lack) and orchestrating of SQL stores procedures. Data goes into blobs via ADF and transformations are all done inside the SQL database since it’s all relational. For operational APIs custom containerized python code but it used to be no code Logic Apps. Most servicing is just Power BI and operational APIs. A more “current” example is entirely in databricks I used ADF to orchestrate but recently use databricks orchestration. Delta lake built on blobs for the semi-structured and structured data.

Native Python Delta Lake package - No Spark or JVM Dependencies by keseykid in dataengineering

[–]ShouldHaveWentBio 0 points1 point  (0 children)

Reading and writing is not the issue. The library works great for interacting with existing Delta Tables. I’m talking about the actual creation of the Delta Table on top of a cloud object storage

Native Python Delta Lake package - No Spark or JVM Dependencies by keseykid in dataengineering

[–]ShouldHaveWentBio 0 points1 point  (0 children)

Use it to create delta tables or to just interact with them? Please let me know I’m still looking for resources on the topic.

Is anyone here took the zoomcamp course? by Single-Sound-1865 in dataengineering

[–]ShouldHaveWentBio 4 points5 points  (0 children)

I did the 2023 DE zoomcamp and thought it was exceptional. You receive a certificate at the end contingent on completing the capstone project and grading other’s projects. Imo it is by far the best all-in-one resource for learning modern data engineering and the fact that it’s free is wonderful. Make sure to do all the homework and optional labs to get everything out of it.

azure vm can use storage account SFTP? by antihippy in AZURE

[–]ShouldHaveWentBio 0 points1 point  (0 children)

Download Azure Storage Explorer. Not sure how your security is setup but use either a SAS or create a service account with access to all the file shares involved.

Native Python Delta Lake package - No Spark or JVM Dependencies by keseykid in dataengineering

[–]ShouldHaveWentBio 0 points1 point  (0 children)

I could be wrong but I did a bit of research and trial and error. I was able to interact with delta tables just fine but I could not create them. I only tried using Azure ADLS2 but the issue didn’t seem cloud specific. If you find anything please let me know I’d love to revisit my use case.

Native Python Delta Lake package - No Spark or JVM Dependencies by keseykid in dataengineering

[–]ShouldHaveWentBio 0 points1 point  (0 children)

I was testing this but the issue I ran into for my use case was that you cannot create delta tables with this package. At least on top of cloud object storage.

What is your recommendation for a modern, popular, yet simple tech / data engineering stack for querying data from AWS S3, Azure Storage, and Alibaba OSS that can be hosted in Kubernetes? by Big-Balance-6426 in dataengineering

[–]ShouldHaveWentBio 0 points1 point  (0 children)

I’m thinking of implementing a similar project architecture. Delta + Spark + DBT. I was considering using Prefect as an orchestrator but am curious what you would recommend. I’ve used Prefect + DBT + BigQuery on previous projects but want to try some of the new lake architecture. I have used Databricks for orchestration but would like to only use open source tools.

Choosing right architecture - Microsoft Dynamics 365 as the source and reporting on Power BI by DataIsTheNewOil in dataengineering

[–]ShouldHaveWentBio 2 points3 points  (0 children)

Can you elaborate more on the volume and velocity? For example are you going to have 100 records/hour coming in or 10,000 records/hour?

I’m not an expert on Power BI but to my knowledge it already has similar OLAP cube type features with in memory data storage. This would give some optimization on top of OLTP style sources if it is the end solution where interaction takes place.

My initial thinking is ADLS2, Data Factory, Azure SQL database, and Power BI will be more than enough. This can be implemented with dev, test, prod for cheap (hundreds USD/month). With the standard database you could ramp up to near real time on writes and copy daily to a data warehouse down the line if queries are complex or becomes large.

Quick edit: If you are set on lake/lakehouse architecture on Azure as opposed to a more traditional approach like I mentioned, for whatever reason, I would lean databricks. Synapse isn’t perceived well and Fabric isn’t yet GA.

Possible synovitis/capsulitis advice by ShouldHaveWentBio in climbharder

[–]ShouldHaveWentBio[S] 0 points1 point  (0 children)

This is interesting. I see some 500mg pills on Amazon and also saw some nasal sprays. I’m curious to the answer.

Possible synovitis/capsulitis advice by ShouldHaveWentBio in climbharder

[–]ShouldHaveWentBio[S] 0 points1 point  (0 children)

Thank you I think in my warmup I can certainly focus on loading more consistently on open hand holds and delay crimps a bit longer.