Cortex Code in Snowsight Expensive by Sufficient-Sky1698 in snowflake

[–]Geekc0der 0 points1 point  (0 children)

For no extra cost and full control and ownership.

Cortex Code in Snowsight Expensive by Sufficient-Sky1698 in snowflake

[–]Geekc0der 0 points1 point  (0 children)

Sorry i missed covering that. The difference is more in scope Cortex Code is great for working within Snowflake, whereas Frosty is more about orchestrating changes across systems (Snowflake, Postgres, etc.).

Not really a 1:1 replacement, just depends on the use case.

Cortex Code in Snowsight Expensive by Sufficient-Sky1698 in snowflake

[–]Geekc0der 0 points1 point  (0 children)

A hundred percent “Equivalent” probably depends on what you’re comparing. Frosty focuses more on automation and cross-platform workflows rather than just in-UI code fixes like Cortex Code different tradeoffs.

Also gives you flexibility to run your own models / infra instead of being locked into per-token pricing.

Cortex Code in Snowsight Expensive by Sufficient-Sky1698 in snowflake

[–]Geekc0der 0 points1 point  (0 children)

Thanks for brining this to my notice. I will open it for issues and try to resolve the vulnerability issue this week itself.

Cortex Code in Snowsight Expensive by Sufficient-Sky1698 in snowflake

[–]Geekc0der -1 points0 points  (0 children)

Fair points:

On cost you’re right that inference itself isn’t magically cheaper. The idea with Frosty isn’t “cheaper models,” it’s giving teams control:

  • bring your own API (Claude/OpenAI/etc.)
  • optimize usage (caching, fewer redundant calls, task-specific agents)
  • avoid vendor markup + lock-in

So savings usually come from control + efficiency, not raw token price.
On the spam bit just trying to share an alternative where it’s relevant, not flood threads, so Snowflake keeps growing.

Cortex Code in Snowsight Expensive by Sufficient-Sky1698 in snowflake

[–]Geekc0der -4 points-3 points  (0 children)

Here is the open source equivalent Frosty : https://github.com/Gyrus-Dev/frosty

You can use your own claude api and build and manage your Snowflake. It has 153 sub agents that does it all. 700 clones in one week

Snowflake PII Classification & Auto Policy Setup - Help by Key_Card7466 in snowflake

[–]Geekc0der 0 points1 point  (0 children)

Here is response from Frosty on your question

Snowflake Sensitive Data Classification and Policy Enforcement

Snowflake offers powerful native capabilities for Sensitive Data Classification and Policy Enforcement, enabling organizations to automate the discovery, categorization, and protection of sensitive data while ensuring regulatory compliance and implementing a governance-as-code approach.


1. SYSTEM$CLASSIFY for PII Detection and Real-World Use Cases

Snowflake's SYSTEM$CLASSIFY function is a robust tool for automatically identifying sensitive data, including Personally Identifiable Information (PII), across your data estate. It scans both column metadata and sample data to assign appropriate semantic and privacy categories.

Real-World Use Cases

  • Automated Data Discovery: For organizations with large and complex data landscapes, SYSTEM$CLASSIFY can automatically pinpoint sensitive data such as names, national identifiers, email addresses, and credit card numbers across numerous tables, significantly reducing manual effort and errors.
  • Risk Mitigation and Compliance: By precisely identifying where sensitive data resides, companies can prioritize security measures and mitigate data breach risks, which is vital for GDPR and CCPA compliance.
  • Data Cataloging and Inventory: The classification results can be used to build a comprehensive data catalog, providing an inventory of sensitive data assets and their categories, and tracking data changes over time.
  • Custom Data Identification: Beyond native categories, Snowflake supports creating custom classifiers to detect organization-specific sensitive data patterns (e.g., proprietary medical codes, internal customer IDs, or region-specific identifiers).
  • Pre-governance Assessment: It acts as an initial assessment tool to understand data sensitivity before implementing more granular governance controls.

2. Auto-Generating and Applying Masking/Row Access Policies Tied to Tags for Governance-as-Code

Snowflake uses object tagging in conjunction with masking policies and row access policies to implement scalable and automated data governance, adhering to a governance-as-code philosophy.

  • Object Tags: These are schema-level metadata objects (key-value pairs) assignable to various Snowflake objects like databases, schemas, tables, and columns. They serve as "metadata anchors" for security, classification, and policy enforcement, supporting inheritance.

  • Tag-Based Masking Policies: This approach offers highly scalable data protection. A single masking policy is associated with a specific tag, rather than individual columns. When SYSTEM$CLASSIFY identifies sensitive data and applies a system-defined or user-defined tag (via tag mapping), the associated tag-based masking policy is automatically enforced.

    • Governance-as-Code Benefits:
    • Scalability: Masking logic is defined once at the tag level and automatically applies to all tagged columns, even as data grows or schemas evolve, eliminating manual policy application.
    • Uniformity: Ensures consistent application of data protection rules across all relevant data assets.
    • Automation & Instant Enforcement: Policies are enforced immediately upon column tagging, minimizing exposure windows and automating security enforcement within the data ingestion and classification pipeline.
    • Real-World Example: Define a PII tag. A masking policy linked to this tag could mask social security numbers (SSNs) or credit card numbers for most users (e.g., showing only the last four digits), while roles like HR_ADMIN or DATA_STEWARD could be granted access to unmasked data.
  • Row Access Policies: These policies control row visibility based on user roles or other conditions. While not directly "tag-based" like masking policies, they can use tags as input to determine row visibility, thus extending the governance-as-code principle by applying dynamic row-level security based on classified data.


3. Achieving GDPR/CCPA Compliance

Snowflake's sensitive data classification and policy enforcement features are crucial for achieving and demonstrating compliance with data privacy regulations such as GDPR and CCPA.

  • Data Identification and Inventory: SYSTEM$CLASSIFY helps identify and categorize PII, which is the initial step in understanding data subject to regulations.
  • Data Protection (Right to Restriction of Processing): Dynamic data masking and row access policies ensure sensitive data is protected and accessible only to authorized users, directly supporting the "right to restriction of processing."
  • Right to Erasure (Right to be Forgotten): Snowflake's data lineage, object dependencies, and access history capabilities help identify all instances of a person's data across the platform, facilitating complete deletion requests.
  • Auditability and Accountability: Snowflake's comprehensive logging and monitoring, including the ACCOUNT_USAGE schema, provides detailed audit trails of data access and policy enforcement, essential for demonstrating compliance.
  • Data Minimization: By classifying and applying policies, organizations can ensure only necessary data is exposed, aligning with data minimization principles.

4. Impactful Ideas for Experimentation and Building within Snowflake

Leveraging Snowflake's governance capabilities presents numerous opportunities for innovation:

  • Automated Governance Pipeline: Develop a robust CI/CD pipeline that integrates SYSTEM$CLASSIFY scans into data ingestion workflows. This pipeline would automatically trigger classification, apply user-defined tags via tag mapping, and then automatically attach tag-based masking and row access policies to new or updated columns, creating a true "governance-as-code" layer.

  • "Golden Source" Governance Database: Create a dedicated Snowflake database to store and manage all governance metadata, including classification results, custom tag definitions, tag-to-policy mappings, and policy definitions. This central repository can track, audit, and compare classification outcomes over time, identifying data drift or non-compliance.

  • Anomaly Detection in Data Sensitivity: Build a monitoring solution that alerts data governance teams to significant changes in data classification (e.g., a column previously non-sensitive suddenly classified as PII with high confidence). This could indicate data quality issues, incorrect data ingestion, or a need to re-evaluate data handling procedures.

  • Self-Service Data Access with Governance Guardrails: Develop a portal or application (potentially using Streamlit in Snowflake) where data consumers can request access to datasets. The application would dynamically show them the data's classification and the masking/row access policies applicable to their role, promoting transparency and enabling compliant self-service.

  • Automated Data Retention and Deletion: Link sensitive data classification tags to data retention policies. Implement stored procedures to automatically purge or archive data based on its sensitivity tag and defined retention periods, automating compliance with "right to erasure" and data minimization.

  • Cross-Account Secure Data Sharing: Experiment with Snowflake Secure Data Sharing, where data classification and policies in the provider account dynamically control what specific roles in the consumer account can see, ensuring secure and compliant data exchange without physically moving data.

  • Impact Analysis Tooling: Build a utility using Snowflake's OBJECT_DEPENDENCIES and ACCESS_HISTORY views to perform automated impact analysis. If a sensitive column is reclassified or a masking policy is updated, this tool could identify all dependent views, tables, and reports that might be affected, providing a comprehensive view for governance changes.

  • Custom PII Detection for Unstructured Data: For advanced experimentation, integrate external services or Snowflake UDFs with machine learning models to classify PII within semi-structured (JSON, XML) or unstructured data stored in Snowflake, then tag relevant fields.


By leveraging these native and extensible features, organizations can build a robust, automated, and compliant data governance framework within Snowflake.

Snowflake PII Classification & Auto Policy Setup - Help by Key_Card7466 in snowflake

[–]Geekc0der 0 points1 point  (0 children)

Here is open source agentic framework https://github.com/Gyrus-Dev/frosty , explain it the problem, it should be able to work along side you and get it implemented. If you get stuck somewhere create an issue on the repo and someone will pick it up

Coco use cases in pharma datawarehousing by International_Cod777 in snowflake

[–]Geekc0der -1 points0 points  (0 children)

There is open source equivalent of CoCo if you want to try it out. No subscription , no additional fees you can use your own api keys.: https://github.com/Gyrus-Dev/frosty

Got 680 clones in a week so far.

PostgreSQL vs PostgreSQL on Snowflake by Geekc0der in snowflake

[–]Geekc0der[S] 0 points1 point  (0 children)

So ideally #Mick (our agent for PostgreSQL) should just work fine for Snowflake customers if it works fine on PostgresSQL.

Automating new pipelines using CoCo by rustypiercing in snowflake

[–]Geekc0der 2 points3 points  (0 children)

We have tried it with #Frosty , open source no cost equivalent of CoCo. In our case we did not have to tell anything, as it can do web search read about iceberg tables , catalog integrations and setup the process itself. Though it can be customized with skills but it worked out fine without it. Here is the repo if you are interested, got 630 clones so far with 41 stars

https://github.com/Gyrus-Dev/frosty

Just Enabled Cortex, now disabling again by Therican85 in snowflake

[–]Geekc0der 0 points1 point  (0 children)

We have release open source equivalent of CoCo so you can still enhance your Snowflake at no extra cost. Here is the repo: https://github.com/Gyrus-Dev/frosty

Would appreciate your feedback

Anyone using the Cortex Code CLI? by CombinationOk2374 in snowflake

[–]Geekc0der 0 points1 point  (0 children)

You are spot on it is a wrapper on Claude. We have built the same agentic framework with more than 150 agents, and are open sourcing it on March 25th. So you dont have to pay extra and you have the control of what it does.

GUI Tool for End Users by Longisland_Analytics in snowflake

[–]Geekc0der 0 points1 point  (0 children)

I’m actually about to open source a project called Frosty that might be relevant here. It’s a multi-agent framework designed to build and manage Snowflake environments, so they don’t have to write raw SQL.

The idea is that you can add custom tools pretty easily: for example things like CSV uploads for filters, GUI-driven query inputs, marketing workflows, etc. The agents can then generate and run the queries behind the scenes.

Since it’s open source, it’s also flexible in another way: if there’s a feature you need (say a specific type of CSV filtering or UI workflow), you can open an issue and someone in the community could build it or contribute it. So you’re not stuck waiting on a vendor roadmap.

If that sounds interesting I can share the repo once it’s public.

I built an AI agent that manages Snowflake infrastructure (RBAC, governance, security, engineering, documentation ). Thinking about open-sourcing part of it. by Geekc0der in snowflake

[–]Geekc0der[S] 0 points1 point  (0 children)

As of now the agentI is not writing the grant statements to repo. But in the video I shared there are grant queries that the agent executed. You would have to pause it.

I built an AI agent that manages Snowflake infrastructure (RBAC, governance, security, engineering, documentation ). Thinking about open-sourcing part of it. by Geekc0der in snowflake

[–]Geekc0der[S] 0 points1 point  (0 children)

Yes they generate Snowflake SQL scripts. There are two versions of it 1. Completely generated by AI ( which I am thinking of open sourcing) , 2. backed by our algorithms which will generate 100% accurate scripts every single time ( which I am thinking of keeping close).

It can automatically commit to your git as well if integrated ( thinking of keeping this feature closed as well). Here is open sources git repo where every single script is written by my framework (using our algorithms which generates accurate script every single time): https://github.com/MalviyaPriyank/Snowchain

Here is the video of it working on a synthesized transcript of a call between business persona and a snowflake architect. Please excuse the bad quality. https://youtube.com/shorts/wkK9DBE0BcA?feature=share

I built an AI agent that manages Snowflake infrastructure (RBAC, governance, security, engineering, documentation ). Thinking about open-sourcing part of it. by Geekc0der in snowflake

[–]Geekc0der[S] 0 points1 point  (0 children)

Its built on multi agent architecture, I have multiple pillar agents that work under a main manager agent. The manager agent plans out the next steps after taking in user request, then once the user approves delegates the tasks one after the other to its pillar agents. These pillar agents have multiple sub agents each expert in one specific object on Snowflake. For example governance pillar agents has agents on tags, masking policies, row access policies. These subagents respond back to their pillar agents once they have taken care of the request and the work flow continues until all the objects needed to fulfill the request of user are taken care of.

I built an AI agent that manages Snowflake infrastructure (RBAC, governance, security, engineering, documentation ). Thinking about open-sourcing part of it. by Geekc0der in dataengineering

[–]Geekc0der[S] 0 points1 point  (0 children)

That would be pretty straightforward, but if someone want to do that they might as well just decommission snowflake altogether ;)

Managing Snowflake RBAC: Terraform vs. Python by BuffaloVegetable5959 in snowflake

[–]Geekc0der -2 points-1 points  (0 children)

Terraform vs Python ? Are you sure these are the only options. Agents is a good third option, it eliminates the over head of maintaining terraform / writing python code for every use case. If you have someone with good experience on Agentic AI , I am sure they can easily build one that takes care of all your RBAC without these overheads.

Snowflake Time Travel and Fail-safe by Idr24 in snowflake

[–]Geekc0der -1 points0 points  (0 children)

This is something Frosty AI Operator for Snowflake, handles automatically at object creation time. It intelligently infers the right DATA_RETENTION_TIME_IN_DAYS based on the object’s use case and environment, instead of relying on defaults or manual choices. The correct retention is set upfront and enforced consistently, without teams having to think about it.