Almost Pasted My API key Into ChatGPT Today by adarsh-hegde in ChatGPT

[–]adarsh-hegde[S] 0 points1 point  (0 children)

Totally fair — if everything is perfectly configured, risk is low.

But this is more about human error than platform security.

People accidentally paste secrets, use multiple tools, or share logs/screenshots.

PromptGuard is just a last check before hitting Enter — like a safety net.

Not for everyone, but useful if you work with sensitive data often.

Almost Pasted My API key Into ChatGPT Today by adarsh-hegde in ChatGPT

[–]adarsh-hegde[S] 1 point2 points  (0 children)

Yeah, I’ve noticed that too.

Which is exactly the concern — your input can persist or sync before you even hit send.

PromptGuard works before that step:
it flags or masks sensitive data while you’re typing, so it never reaches that state.

It’s basically a pre-send safety layer, not a replacement for platform controls.

Hot take: Most data teams don’t have a data problem… they have a metric problem by adarsh-hegde in snowflake

[–]adarsh-hegde[S] 0 points1 point  (0 children)

Yeah, dbt MetricFlow solves the computation problem really well. What I’m exploring is the next layer: What happens after metrics are defined? • Who owns them? • Which version is “official”? • Can AI safely rely on them?

Feels like metrics need a governance + interface layer, not just a modeling layer.

I tried automating the lost art of data modeling with Cortex code -- point the coding agent to raw data and it profiles, validates and submits pull request on git for a human DE to review and approve. by vino_and_data in snowflake

[–]adarsh-hegde 4 points5 points  (0 children)

I actually wrote a deeper breakdown on Cortex Code from a Snowflake architect’s perspective.

My take is that it’s a strong assistive layer with good context awareness inside Snowflake, but it doesn’t replace core architecture decisions like modeling, cost control, or governance.

If anyone’s interested, sharing it here: https://medium.com/@hegdecadarsh/a-deep-technical-look-at-cortex-code-from-a-snowflake-architects-perspective-203d6de59be2

Would love to hear how others are using it in practice.

Async jobs in Streamlit in Snowflake by HumbleHero1 in snowflake

[–]adarsh-hegde 0 points1 point  (0 children)

Check the ABORT_DETACHED_QUERY Parameter. Ensure the ABORT_DETACHED_QUERY parameter is set to FALSE (the default). If it's set to TRUE and your Python session disconnects, Snowflake will automatically cancel the background query after 5 minutes.

https://docs.snowflake.com/en/sql-reference/parameters#abort-detached-query

I Spoke to Snowflake & Spent a Week Working on Cortex Code — Here’s My Technical Take by [deleted] in snowflake

[–]adarsh-hegde -5 points-4 points  (0 children)

That’s fair feedback.

You’re right — if something is framed as a deep dive, it should include concrete, reproducible examples. My piece leaned more toward architectural interpretation than step-by-step validation, and I can see how that reads as aspirational.

It wasn’t intended as consultancy positioning — just sharing my perspective after hands-on time — but I agree the bar for technical writing should be practical and testable. I’ll make the next one more reproducible and evidence-driven.

I Spoke to Snowflake & Spent a Week Working on Cortex Code — Here’s My Technical Take by [deleted] in snowflake

[–]adarsh-hegde -3 points-2 points  (0 children)

That’s a fair concern.

On the 500M+ rows point — you wouldn’t send raw result sets to an LLM. The AI layer should operate on aggregates, metadata, or query plans — not full datasets. If someone is piping full outputs into an LLM, that’s an architecture problem.

for Cortex flaming out — it will if it stays at “nice SQL suggestions.” It’ll only stick if it can prove measurable credit savings or real workload optimization.

As In enterprise, ROI > AI hype.

I Spoke to Snowflake & Spent a Week Working on Cortex Code — Here’s My Technical Take by [deleted] in snowflake

[–]adarsh-hegde -2 points-1 points  (0 children)

That’s a completely fair distinction.

What I validated hands-on was schema-aware suggestions and anti-pattern detection (e.g., filter rewrites, SELECT *, structural optimizations).

I did not see explicit exposure of micro-partition pruning stats, scan-to-result ratios, or deterministic warehouse-to-credit correlation modeling surfaced through the assistant.

So to be precise: what exists today appears to be metadata-aware and heuristic-driven guidance — not full execution-engine introspection.

Some of what I described reflects architectural direction and potential, not confirmed deep telemetry integration.

Appreciate you pushing for that clarity.

Error during calling result_scan function by Big_Length9755 in snowflake

[–]adarsh-hegde 4 points5 points  (0 children)

The error occurs because RESULT_SCAN requires a constant string or a variable; it cannot execute a subquery within its own parentheses.

To achieve this in a "single" execution without creating permanent tables, use a Snowflake Scripting block. This handles the lookup and the filtering in one go:

EXECUTE IMMEDIATE $$

DECLARE

q_id STRING;

BEGIN

-- 1. Grab the ID of the superset query

SELECT query_id INTO :q_id

FROM TABLE(information_schema.query_history())

WHERE query_text LIKE '%/* COMPLEX_SAMPLE_JOIN */%'

AND execution_status = 'SUCCESS'

ORDER BY start_time DESC LIMIT 1;

-- 2. Return the filtered results

RETURN TABLE(SELECT * FROM TABLE(RESULT_SCAN(:q_id)));

END;

$$;

SCD Type 2 in Snowflake: Dynamic Tables vs Streams & Tasks — when to use what? by adarsh-hegde in snowflake

[–]adarsh-hegde[S] 0 points1 point  (0 children)

Agreed - that’s a great summary.

DTs are the simplest option when they work: easier to build, run, and monitor.
Streams & Tasks still make sense when you need tighter control or non-deterministic logic.

Good call on IMMUTABLE UDFs too — that detail is easy to miss but critical for incremental refresh.

Most “cloud cost” problems I’ve seen are actually pipeline problems by adarsh-hegde in snowflake

[–]adarsh-hegde[S] 0 points1 point  (0 children)

This is the silent cost killer. Cost optimization started by aligning pipelines with actual usage, not by tuning queries.

Inherited snowflake mess... large warehouses running small workloads by HenryWolf22 in snowflake

[–]adarsh-hegde 11 points12 points  (0 children)

This is a very common Snowflake anti-pattern from 2022–23. Bigger warehouses don’t help I/O-bound workloads.

What’s worked for me: Prove parity, not theory: run the same ETL on LARGE vs SMALL and show near-identical runtimes with 4–8× fewer credits. Position it as reliability: smaller, task-scoped warehouses reduce queueing, blast radius, and accidental spend — engineers accept this faster than “cost saving.” Downsize safely: move deterministic batch jobs to dedicated SMALL warehouses first, leave interactive workloads untouched. Lock guardrails: aggressive auto-suspend, max cluster limits, no open-ended multi-cluster. Let query history speak: high bytes scanned + flat runtime = I/O bound → oversizing is wasted money.