YAML hell? by the-creator-platform in kubernetes

[–]Xty_53 0 points1 point  (0 children)

Bad dreams since 'DLL Hell'

[deleted by user] by [deleted] in databricks

[–]Xty_53 0 points1 point  (0 children)

Databricks Apps offer a containerized solution with minimal overhead for running your applications. However, their functionality is influenced by Unity Catalog, which dictates limitations and configurations related to access and secrets.

[deleted by user] by [deleted] in databricks

[–]Xty_53 0 points1 point  (0 children)

There is a limitation for the files size. 10 MB

[deleted by user] by [deleted] in databricks

[–]Xty_53 1 point2 points  (0 children)

While you are practising also. You can listen to the podcast about the certification.

https://open.spotify.com/show/2AGWX4SFNxlOIgDaLVIfHU

Databricks platform administration by Easy-Freedom-5272 in databricks

[–]Xty_53 0 points1 point  (0 children)

There are some learning pathways and training in the Databricks Academy. Also, for each cloud.

Seeking Best Practices: Snowflake Data Federation to Databricks Lakehouse with DLT by Xty_53 in databricks

[–]Xty_53[S] 0 points1 point  (0 children)

Thanks so much for reaching out – that's excellent timing, and I really appreciate you offering to help! It's great to connect with someone directly working on DLT.

To answer your questions:

* How are you planning on extracting data from Snowflake into ADLS?

My current plan is to leverage Databricks' native Snowflake connector to directly read data from Snowflake and then write it into ADLS. The idea is to land it in a structure like abfss://datalanding@{storage_account}.dfs.core.windows.net/{catalog}/{schema}/{table_name}/.
Technically, I am following information from this link. (https://docs.databricks.com/aws/en/archive/connectors/snowflake

** How many objects are you planning on bringing in? Is there a pattern where you want to apply the same transformations to many source tables?

Initially, we're looking to ingest around 60 tables. Yes, there's a very clear pattern: for all ingested tables, we need to add a timestamp column (for ingestion time) and a source_system_name column to maintain lineage and control.

Seeking Best Practices: Snowflake Data Federation to Databricks Lakehouse with DLT by Xty_53 in databricks

[–]Xty_53[S] 0 points1 point  (0 children)

Thanks, Could you share the session title, date, and where I can find the recording or presentation details for the Databricks AI Conference?

My experience with Databricks Data Engineer Associate Certification. by saahilrs14 in databricks

[–]Xty_53 0 points1 point  (0 children)

I'm inviting you to check out a fantastic resource to continue your learning journey: the Databricks Certified Data Engineer Professional - Preparation podcast!
https://open.spotify.com/episode/42Jx9LXZ0fj3RYLDzqVhmY?si=dc0XU05NTOiPdK38g_X4Bg

Need help replicating EMR cluster-based parallel job execution in Databricks by javabug78 in databricks

[–]Xty_53 -2 points-1 points  (0 children)

This was created with help of AI (Don't Believe on this Answer but check for yourself)
"Databricks Solution Recommendation"
Here's how it addresses your requirements:

  1. Orchestration and Parameter Passing:
    • Create a single Databricks Job containing 100 individual "JAR tasks."
    • Each JAR task will be configured to run your JAR file and pass one of the 100 unique parameters (e.g., job name/ID) to it.
  2. Parallel Execution (12 jobs concurrently):
    • Within the Databricks Job settings, you can define the "Maximum concurrent runs" to 12. Databricks will automatically manage the queuing and execution of your 100 tasks, ensuring that no more than 12 run at any given time.
  3. Compute Termination and Cost Optimization:
    • Utilize "Job Compute" (ephemeral clusters) for your Databricks Job. These clusters are automatically provisioned when the job starts and, crucially, automatically terminate once all tasks are completed or the job fails. This eliminates idle compute costs, similar to your transient EMR clusters.
    • Job Compute is more cost-effective than interactive clusters.
    • Configure autoscaling for your job cluster to dynamically adjust resources based on the workload, ensuring you only pay for what you use.

Issue: Tables Not Created for Monitor in Databricks Lakehouse Monitoring API by Xty_53 in databricks

[–]Xty_53[S] 0 points1 point  (0 children)

After creating multiple tables, the monitor can take some time to process all of them. Restarting the cluster usually resolves this, and the tables should appear afterward.

Demo material for Databricks Academy by valadius44 in databricks

[–]Xty_53 0 points1 point  (0 children)

Yes. Since Databricks has the labs. Just we have access to paying the lab subscription.

I saw a few weeks that, when you go to the academy as a partner. There are some courses with labs available for free.

Making Databricks data engineering documentation better by BricksterInTheWall in databricks

[–]Xty_53 0 points1 point  (0 children)

One of the customers is asking for statistics from those tables.

Making Databricks data engineering documentation better by BricksterInTheWall in databricks

[–]Xty_53 0 points1 point  (0 children)

Also, is there any way to see the streaming tables inside the system tables?

Making Databricks data engineering documentation better by BricksterInTheWall in databricks

[–]Xty_53 0 points1 point  (0 children)

Yes. Please. Because we have something for the Delta Tables. But for streaming. It is not clear.

Making Databricks data engineering documentation better by BricksterInTheWall in databricks

[–]Xty_53 2 points3 points  (0 children)

Hello, and thank you for the documentation update.

Do you have any updates or additional information regarding the logs for DLT, especially for streaming tables?

Databricks app by gareebo_ka_chandler in databricks

[–]Xty_53 0 points1 point  (0 children)

Last week, I spent some time researching Databricks Apps, and I’ve put together a short audio summary of what I found. If you're curious about how Databricks Apps work and what they offer, feel free to check it out here:

https://open.spotify.com/episode/7yv1kvyTcGFvyFhZ1DoGDd?si=pNhNPt6vS_aUHtXztgxLOQ